[Intel-gfx] [PATCH] gpu/drm/i915: globally replace dev_priv with i915

Rodrigo Vivi rodrigo.vivi at intel.com
Thu Jun 13 16:23:49 UTC 2019


On Thu, Jun 13, 2019 at 04:12:16PM +0100, Tvrtko Ursulin wrote:
> 
> On 12/06/2019 17:29, Lucas De Marchi wrote:
> > We are slowly converting dev_priv to i915 everywhere, spread into
> > smaller series. While this is good to avoid unrelated breakages to other
> > inflight patches, it's bad because inflight patches on nearby paths keep
> > breaking. Paired with other code moves and refactores this is becoming a
> > nightmare.
> > 
> > Now that I915_{READ,WRITE} are gone and implicit access to dev_priv no
> 
> Who removed I915_READ/WRITE? In fact I see some in this patch. Well colour
> me confused.. how did the patch compile?

as soon as this becomes true I agree with Lucas that the big sed is the
best solution.

> 
> Plus it's not I915_READ/WRITE - some display macros have hidden dependency
> on implicit dev_priv as well.
> 
> > longer exists we can simply sed all files and be done with the
> > conversion. This was generated with the following commands with no
> > additional fixups:
> > 
> > 	git ls-files --full-name -z -- drivers/gpu/drm/i915/ | \
> > 		xargs -0 sed -i 's/\bdev_priv\b/i915/g'
> > 
> > Any pending series can apply the same search and replace when rebasing.
> 
> I am in two minds. Converting piece meal is easier for some use cases, and
> converting all at once is easier for some others. Don't know.
> 
> Regards,
> 
> Tvrtko
> 
> > 
> > Signed-off-by: Lucas De Marchi <lucas.demarchi at intel.com>
> > ---
> >   drivers/gpu/drm/i915/dvo_ch7017.c             |    6 +-
> >   drivers/gpu/drm/i915/dvo_ch7xxx.c             |   10 +-
> >   drivers/gpu/drm/i915/dvo_ivch.c               |   14 +-
> >   drivers/gpu/drm/i915/dvo_ns2501.c             |   14 +-
> >   drivers/gpu/drm/i915/dvo_sil164.c             |   10 +-
> >   drivers/gpu/drm/i915/dvo_tfp410.c             |   10 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_context.c   |   48 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_context.h   |    6 +-
> >   .../gpu/drm/i915/gem/i915_gem_execbuffer.c    |    4 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_object.c    |    4 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_shmem.c     |    4 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_stolen.c    |  168 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_tiling.c    |    8 +-
> >   drivers/gpu/drm/i915/gem/i915_gem_userptr.c   |   40 +-
> >   .../gpu/drm/i915/gem/selftests/huge_pages.c   |   48 +-
> >   .../drm/i915/gem/selftests/i915_gem_context.c |    6 +-
> >   drivers/gpu/drm/i915/gt/intel_engine.h        |    2 +-
> >   drivers/gpu/drm/i915/gt/intel_engine_cs.c     |   66 +-
> >   drivers/gpu/drm/i915/gt/intel_hangcheck.c     |   36 +-
> >   drivers/gpu/drm/i915/gt/intel_mocs.c          |   24 +-
> >   drivers/gpu/drm/i915/gt/intel_mocs.h          |    2 +-
> >   drivers/gpu/drm/i915/gt/intel_reset.c         |    6 +-
> >   drivers/gpu/drm/i915/gt/intel_ringbuffer.c    |   52 +-
> >   drivers/gpu/drm/i915/gvt/aperture_gm.c        |   60 +-
> >   drivers/gpu/drm/i915/gvt/cfg_space.c          |    4 +-
> >   drivers/gpu/drm/i915/gvt/cmd_parser.c         |   34 +-
> >   drivers/gpu/drm/i915/gvt/debugfs.c            |   12 +-
> >   drivers/gpu/drm/i915/gvt/display.c            |   48 +-
> >   drivers/gpu/drm/i915/gvt/dmabuf.c             |   12 +-
> >   drivers/gpu/drm/i915/gvt/edid.c               |    6 +-
> >   drivers/gpu/drm/i915/gvt/execlist.c           |   14 +-
> >   drivers/gpu/drm/i915/gvt/fb_decoder.c         |   12 +-
> >   drivers/gpu/drm/i915/gvt/firmware.c           |   16 +-
> >   drivers/gpu/drm/i915/gvt/gtt.c                |   44 +-
> >   drivers/gpu/drm/i915/gvt/gvt.c                |   26 +-
> >   drivers/gpu/drm/i915/gvt/gvt.h                |   20 +-
> >   drivers/gpu/drm/i915/gvt/handlers.c           |   92 +-
> >   drivers/gpu/drm/i915/gvt/interrupt.c          |    6 +-
> >   drivers/gpu/drm/i915/gvt/kvmgt.c              |    6 +-
> >   drivers/gpu/drm/i915/gvt/mmio.c               |    2 +-
> >   drivers/gpu/drm/i915/gvt/mmio_context.c       |   38 +-
> >   drivers/gpu/drm/i915/gvt/sched_policy.c       |   12 +-
> >   drivers/gpu/drm/i915/gvt/scheduler.c          |   72 +-
> >   drivers/gpu/drm/i915/gvt/vgpu.c               |    6 +-
> >   drivers/gpu/drm/i915/i915_cmd_parser.c        |    6 +-
> >   drivers/gpu/drm/i915/i915_debugfs.c           |  952 +++---
> >   drivers/gpu/drm/i915/i915_debugfs.h           |    4 +-
> >   drivers/gpu/drm/i915/i915_drv.c               | 1092 +++---
> >   drivers/gpu/drm/i915/i915_drv.h               |  590 ++--
> >   drivers/gpu/drm/i915/i915_gem.c               |  356 +-
> >   drivers/gpu/drm/i915/i915_gem_evict.c         |    6 +-
> >   drivers/gpu/drm/i915/i915_gem_gtt.c           |  214 +-
> >   drivers/gpu/drm/i915/i915_gem_gtt.h           |   20 +-
> >   drivers/gpu/drm/i915/i915_gpu_error.c         |   44 +-
> >   drivers/gpu/drm/i915/i915_gpu_error.h         |    6 +-
> >   drivers/gpu/drm/i915/i915_irq.c               | 1808 +++++-----
> >   drivers/gpu/drm/i915/i915_irq.h               |   86 +-
> >   drivers/gpu/drm/i915/i915_memcpy.c            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_bdw.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_bdw.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_bxt.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_bxt.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_cflgt2.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_cflgt2.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_cflgt3.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_cflgt3.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_chv.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_chv.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_cnl.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_cnl.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_glk.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_glk.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_hsw.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_hsw.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_icl.c            |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_icl.h            |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_kblgt2.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_kblgt2.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_kblgt3.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_kblgt3.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt2.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt2.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt3.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt3.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt4.c         |   32 +-
> >   drivers/gpu/drm/i915/i915_oa_sklgt4.h         |    2 +-
> >   drivers/gpu/drm/i915/i915_perf.c              |  814 ++---
> >   drivers/gpu/drm/i915/i915_pmu.c               |   42 +-
> >   drivers/gpu/drm/i915/i915_query.c             |   10 +-
> >   drivers/gpu/drm/i915/i915_reg.h               |  420 +--
> >   drivers/gpu/drm/i915/i915_suspend.c           |  106 +-
> >   drivers/gpu/drm/i915/i915_sysfs.c             |  172 +-
> >   drivers/gpu/drm/i915/i915_trace.h             |   46 +-
> >   drivers/gpu/drm/i915/i915_vgpu.c              |   30 +-
> >   drivers/gpu/drm/i915/i915_vgpu.h              |   16 +-
> >   drivers/gpu/drm/i915/i915_vma.c               |    4 +-
> >   drivers/gpu/drm/i915/icl_dsi.c                |  122 +-
> >   drivers/gpu/drm/i915/intel_atomic.c           |   28 +-
> >   drivers/gpu/drm/i915/intel_atomic.h           |    2 +-
> >   drivers/gpu/drm/i915/intel_audio.c            |  222 +-
> >   drivers/gpu/drm/i915/intel_audio.h            |    6 +-
> >   drivers/gpu/drm/i915/intel_bios.c             |  552 ++--
> >   drivers/gpu/drm/i915/intel_bios.h             |   18 +-
> >   drivers/gpu/drm/i915/intel_bw.c               |   68 +-
> >   drivers/gpu/drm/i915/intel_bw.h               |    8 +-
> >   drivers/gpu/drm/i915/intel_cdclk.c            |  882 ++---
> >   drivers/gpu/drm/i915/intel_cdclk.h            |   14 +-
> >   drivers/gpu/drm/i915/intel_color.c            |  176 +-
> >   drivers/gpu/drm/i915/intel_combo_phy.c        |   64 +-
> >   drivers/gpu/drm/i915/intel_combo_phy.h        |    6 +-
> >   drivers/gpu/drm/i915/intel_connector.c        |   12 +-
> >   drivers/gpu/drm/i915/intel_crt.c              |  172 +-
> >   drivers/gpu/drm/i915/intel_crt.h              |    4 +-
> >   drivers/gpu/drm/i915/intel_csr.c              |  144 +-
> >   drivers/gpu/drm/i915/intel_ddi.c              |  592 ++--
> >   drivers/gpu/drm/i915/intel_ddi.h              |    6 +-
> >   drivers/gpu/drm/i915/intel_device_info.c      |  164 +-
> >   drivers/gpu/drm/i915/intel_device_info.h      |    6 +-
> >   drivers/gpu/drm/i915/intel_display.c          | 2916 ++++++++---------
> >   drivers/gpu/drm/i915/intel_display.h          |    6 +-
> >   drivers/gpu/drm/i915/intel_display_power.c    | 1020 +++---
> >   drivers/gpu/drm/i915/intel_display_power.h    |   62 +-
> >   drivers/gpu/drm/i915/intel_dp.c               |  822 ++---
> >   drivers/gpu/drm/i915/intel_dp.h               |   18 +-
> >   drivers/gpu/drm/i915/intel_dp_aux_backlight.c |    4 +-
> >   drivers/gpu/drm/i915/intel_dp_mst.c           |   32 +-
> >   drivers/gpu/drm/i915/intel_dpio_phy.c         |  276 +-
> >   drivers/gpu/drm/i915/intel_dpio_phy.h         |   12 +-
> >   drivers/gpu/drm/i915/intel_dpll_mgr.c         |  356 +-
> >   drivers/gpu/drm/i915/intel_dpll_mgr.h         |   20 +-
> >   drivers/gpu/drm/i915/intel_drv.h              |   72 +-
> >   drivers/gpu/drm/i915/intel_dsi.c              |    6 +-
> >   drivers/gpu/drm/i915/intel_dsi.h              |    6 +-
> >   .../gpu/drm/i915/intel_dsi_dcs_backlight.c    |    4 +-
> >   drivers/gpu/drm/i915/intel_dsi_vbt.c          |   76 +-
> >   drivers/gpu/drm/i915/intel_dvo.c              |   32 +-
> >   drivers/gpu/drm/i915/intel_dvo.h              |    2 +-
> >   drivers/gpu/drm/i915/intel_dvo_dev.h          |    2 +-
> >   drivers/gpu/drm/i915/intel_fbc.c              |  410 +--
> >   drivers/gpu/drm/i915/intel_fbc.h              |   20 +-
> >   drivers/gpu/drm/i915/intel_fbdev.c            |   46 +-
> >   drivers/gpu/drm/i915/intel_fbdev.h            |    8 +-
> >   drivers/gpu/drm/i915/intel_fifo_underrun.c    |  160 +-
> >   drivers/gpu/drm/i915/intel_fifo_underrun.h    |   12 +-
> >   drivers/gpu/drm/i915/intel_frontbuffer.c      |   82 +-
> >   drivers/gpu/drm/i915/intel_frontbuffer.h      |    6 +-
> >   drivers/gpu/drm/i915/intel_gmbus.c            |  230 +-
> >   drivers/gpu/drm/i915/intel_gmbus.h            |   10 +-
> >   drivers/gpu/drm/i915/intel_guc.c              |   48 +-
> >   drivers/gpu/drm/i915/intel_guc_ads.c          |   12 +-
> >   drivers/gpu/drm/i915/intel_guc_fw.c           |   20 +-
> >   drivers/gpu/drm/i915/intel_guc_log.c          |   18 +-
> >   drivers/gpu/drm/i915/intel_guc_submission.c   |   56 +-
> >   drivers/gpu/drm/i915/intel_gvt.c              |   36 +-
> >   drivers/gpu/drm/i915/intel_gvt.h              |   16 +-
> >   drivers/gpu/drm/i915/intel_hdcp.c             |  286 +-
> >   drivers/gpu/drm/i915/intel_hdcp.h             |    6 +-
> >   drivers/gpu/drm/i915/intel_hdmi.c             |  316 +-
> >   drivers/gpu/drm/i915/intel_hdmi.h             |    2 +-
> >   drivers/gpu/drm/i915/intel_hotplug.c          |  210 +-
> >   drivers/gpu/drm/i915/intel_hotplug.h          |   16 +-
> >   drivers/gpu/drm/i915/intel_huc.c              |    6 +-
> >   drivers/gpu/drm/i915/intel_huc_fw.c           |   18 +-
> >   drivers/gpu/drm/i915/intel_lpe_audio.c        |   96 +-
> >   drivers/gpu/drm/i915/intel_lpe_audio.h        |    8 +-
> >   drivers/gpu/drm/i915/intel_lspcon.c           |    4 +-
> >   drivers/gpu/drm/i915/intel_lvds.c             |   96 +-
> >   drivers/gpu/drm/i915/intel_lvds.h             |    8 +-
> >   drivers/gpu/drm/i915/intel_opregion.c         |  114 +-
> >   drivers/gpu/drm/i915/intel_opregion.h         |   28 +-
> >   drivers/gpu/drm/i915/intel_overlay.c          |   98 +-
> >   drivers/gpu/drm/i915/intel_overlay.h          |    8 +-
> >   drivers/gpu/drm/i915/intel_panel.c            |  238 +-
> >   drivers/gpu/drm/i915/intel_pipe_crc.c         |  120 +-
> >   drivers/gpu/drm/i915/intel_pipe_crc.h         |    4 +-
> >   drivers/gpu/drm/i915/intel_pm.c               | 2300 ++++++-------
> >   drivers/gpu/drm/i915/intel_pm.h               |   74 +-
> >   drivers/gpu/drm/i915/intel_psr.c              |  394 +--
> >   drivers/gpu/drm/i915/intel_psr.h              |   14 +-
> >   drivers/gpu/drm/i915/intel_quirks.h           |    2 +-
> >   drivers/gpu/drm/i915/intel_runtime_pm.h       |    8 +-
> >   drivers/gpu/drm/i915/intel_sdvo.c             |  128 +-
> >   drivers/gpu/drm/i915/intel_sdvo.h             |    4 +-
> >   drivers/gpu/drm/i915/intel_sprite.c           |  218 +-
> >   drivers/gpu/drm/i915/intel_sprite.h           |    8 +-
> >   drivers/gpu/drm/i915/intel_tv.c               |   72 +-
> >   drivers/gpu/drm/i915/intel_tv.h               |    2 +-
> >   drivers/gpu/drm/i915/intel_uc.c               |    4 +-
> >   drivers/gpu/drm/i915/intel_uc.h               |   22 +-
> >   drivers/gpu/drm/i915/intel_uc_fw.c            |   12 +-
> >   drivers/gpu/drm/i915/intel_uc_fw.h            |    2 +-
> >   drivers/gpu/drm/i915/intel_uncore.c           |   26 +-
> >   drivers/gpu/drm/i915/intel_uncore.h           |    4 +-
> >   drivers/gpu/drm/i915/intel_vbt_defs.h         |    2 +-
> >   drivers/gpu/drm/i915/intel_vdsc.c             |   10 +-
> >   drivers/gpu/drm/i915/intel_wopcm.c            |   14 +-
> >   drivers/gpu/drm/i915/selftests/i915_gem_gtt.c |   26 +-
> >   drivers/gpu/drm/i915/selftests/intel_guc.c    |   46 +-
> >   drivers/gpu/drm/i915/selftests/intel_uncore.c |   10 +-
> >   drivers/gpu/drm/i915/vlv_dsi.c                |  222 +-
> >   drivers/gpu/drm/i915/vlv_dsi_pll.c            |   80 +-
> >   201 files changed, 12020 insertions(+), 12020 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/i915/dvo_ch7017.c b/drivers/gpu/drm/i915/dvo_ch7017.c
> > index 602380fe74f3..df670c8d5c31 100644
> > --- a/drivers/gpu/drm/i915/dvo_ch7017.c
> > +++ b/drivers/gpu/drm/i915/dvo_ch7017.c
> > @@ -210,7 +210,7 @@ static bool ch7017_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = priv;
> > +	dvo->i915 = priv;
> >   	if (!ch7017_read(dvo, CH7017_DEVICE_ID, &val))
> >   		goto fail;
> > @@ -395,11 +395,11 @@ do {							\
> >   static void ch7017_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct ch7017_priv *priv = dvo->dev_priv;
> > +	struct ch7017_priv *priv = dvo->i915;
> >   	if (priv) {
> >   		kfree(priv);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/dvo_ch7xxx.c b/drivers/gpu/drm/i915/dvo_ch7xxx.c
> > index e070bebee7b5..e076801d032f 100644
> > --- a/drivers/gpu/drm/i915/dvo_ch7xxx.c
> > +++ b/drivers/gpu/drm/i915/dvo_ch7xxx.c
> > @@ -135,7 +135,7 @@ static char *ch7xxx_get_did(u8 did)
> >   /** Reads an 8 bit register */
> >   static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   {
> > -	struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> > +	struct ch7xxx_priv *ch7xxx = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	u8 in_buf[2];
> > @@ -173,7 +173,7 @@ static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   /** Writes an 8 bit register */
> >   static bool ch7xxx_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> >   {
> > -	struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> > +	struct ch7xxx_priv *ch7xxx = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	struct i2c_msg msg = {
> > @@ -210,7 +210,7 @@ static bool ch7xxx_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = ch7xxx;
> > +	dvo->i915 = ch7xxx;
> >   	ch7xxx->quiet = true;
> >   	if (!ch7xxx_readb(dvo, CH7xxx_REG_VID, &vendor))
> > @@ -347,11 +347,11 @@ static void ch7xxx_dump_regs(struct intel_dvo_device *dvo)
> >   static void ch7xxx_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct ch7xxx_priv *ch7xxx = dvo->dev_priv;
> > +	struct ch7xxx_priv *ch7xxx = dvo->i915;
> >   	if (ch7xxx) {
> >   		kfree(ch7xxx);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/dvo_ivch.c b/drivers/gpu/drm/i915/dvo_ivch.c
> > index 09dba35f3ffa..c99c16211567 100644
> > --- a/drivers/gpu/drm/i915/dvo_ivch.c
> > +++ b/drivers/gpu/drm/i915/dvo_ivch.c
> > @@ -191,7 +191,7 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo);
> >    */
> >   static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data)
> >   {
> > -	struct ivch_priv *priv = dvo->dev_priv;
> > +	struct ivch_priv *priv = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[1];
> >   	u8 in_buf[2];
> > @@ -234,7 +234,7 @@ static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data)
> >   /* Writes a 16-bit register on the ivch */
> >   static bool ivch_write(struct intel_dvo_device *dvo, int addr, u16 data)
> >   {
> > -	struct ivch_priv *priv = dvo->dev_priv;
> > +	struct ivch_priv *priv = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[3];
> >   	struct i2c_msg msg = {
> > @@ -272,7 +272,7 @@ static bool ivch_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = priv;
> > +	dvo->i915 = priv;
> >   	priv->quiet = true;
> >   	if (!ivch_read(dvo, VR00, &temp))
> > @@ -328,7 +328,7 @@ static enum drm_mode_status ivch_mode_valid(struct intel_dvo_device *dvo,
> >    */
> >   static void ivch_reset(struct intel_dvo_device *dvo)
> >   {
> > -	struct ivch_priv *priv = dvo->dev_priv;
> > +	struct ivch_priv *priv = dvo->i915;
> >   	int i;
> >   	DRM_DEBUG_KMS("Resetting the IVCH registers\n");
> > @@ -398,7 +398,7 @@ static void ivch_mode_set(struct intel_dvo_device *dvo,
> >   			  const struct drm_display_mode *mode,
> >   			  const struct drm_display_mode *adjusted_mode)
> >   {
> > -	struct ivch_priv *priv = dvo->dev_priv;
> > +	struct ivch_priv *priv = dvo->i915;
> >   	u16 vr40 = 0;
> >   	u16 vr01 = 0;
> >   	u16 vr10;
> > @@ -483,11 +483,11 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo)
> >   static void ivch_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct ivch_priv *priv = dvo->dev_priv;
> > +	struct ivch_priv *priv = dvo->i915;
> >   	if (priv) {
> >   		kfree(priv);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/dvo_ns2501.c b/drivers/gpu/drm/i915/dvo_ns2501.c
> > index c83a5d88d62b..af04fcaa98b8 100644
> > --- a/drivers/gpu/drm/i915/dvo_ns2501.c
> > +++ b/drivers/gpu/drm/i915/dvo_ns2501.c
> > @@ -392,7 +392,7 @@ struct ns2501_priv {
> >   */
> >   static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   {
> > -	struct ns2501_priv *ns = dvo->dev_priv;
> > +	struct ns2501_priv *ns = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	u8 in_buf[2];
> > @@ -437,7 +437,7 @@ static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   */
> >   static bool ns2501_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> >   {
> > -	struct ns2501_priv *ns = dvo->dev_priv;
> > +	struct ns2501_priv *ns = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> > @@ -481,7 +481,7 @@ static bool ns2501_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = ns;
> > +	dvo->i915 = ns;
> >   	ns->quiet = true;
> >   	if (!ns2501_readb(dvo, NS2501_VID_LO, &ch))
> > @@ -551,7 +551,7 @@ static void ns2501_mode_set(struct intel_dvo_device *dvo,
> >   			    const struct drm_display_mode *adjusted_mode)
> >   {
> >   	const struct ns2501_configuration *conf;
> > -	struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv);
> > +	struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915);
> >   	int mode_idx, i;
> >   	DRM_DEBUG_KMS
> > @@ -655,7 +655,7 @@ static bool ns2501_get_hw_state(struct intel_dvo_device *dvo)
> >   /* set the NS2501 power state */
> >   static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable)
> >   {
> > -	struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv);
> > +	struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915);
> >   	DRM_DEBUG_KMS("Trying set the dpms of the DVO to %i\n", enable);
> > @@ -691,11 +691,11 @@ static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable)
> >   static void ns2501_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct ns2501_priv *ns = dvo->dev_priv;
> > +	struct ns2501_priv *ns = dvo->i915;
> >   	if (ns) {
> >   		kfree(ns);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/dvo_sil164.c b/drivers/gpu/drm/i915/dvo_sil164.c
> > index 04698eaeb632..a452dcba179c 100644
> > --- a/drivers/gpu/drm/i915/dvo_sil164.c
> > +++ b/drivers/gpu/drm/i915/dvo_sil164.c
> > @@ -68,7 +68,7 @@ struct sil164_priv {
> >   static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   {
> > -	struct sil164_priv *sil = dvo->dev_priv;
> > +	struct sil164_priv *sil = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	u8 in_buf[2];
> > @@ -105,7 +105,7 @@ static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   static bool sil164_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> >   {
> > -	struct sil164_priv *sil = dvo->dev_priv;
> > +	struct sil164_priv *sil = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	struct i2c_msg msg = {
> > @@ -142,7 +142,7 @@ static bool sil164_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = sil;
> > +	dvo->i915 = sil;
> >   	sil->quiet = true;
> >   	if (!sil164_readb(dvo, SIL164_VID_LO, &ch))
> > @@ -260,11 +260,11 @@ static void sil164_dump_regs(struct intel_dvo_device *dvo)
> >   static void sil164_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct sil164_priv *sil = dvo->dev_priv;
> > +	struct sil164_priv *sil = dvo->i915;
> >   	if (sil) {
> >   		kfree(sil);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/dvo_tfp410.c b/drivers/gpu/drm/i915/dvo_tfp410.c
> > index 623114ee73cd..8ffed8914b49 100644
> > --- a/drivers/gpu/drm/i915/dvo_tfp410.c
> > +++ b/drivers/gpu/drm/i915/dvo_tfp410.c
> > @@ -93,7 +93,7 @@ struct tfp410_priv {
> >   static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   {
> > -	struct tfp410_priv *tfp = dvo->dev_priv;
> > +	struct tfp410_priv *tfp = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	u8 in_buf[2];
> > @@ -130,7 +130,7 @@ static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch)
> >   static bool tfp410_writeb(struct intel_dvo_device *dvo, int addr, u8 ch)
> >   {
> > -	struct tfp410_priv *tfp = dvo->dev_priv;
> > +	struct tfp410_priv *tfp = dvo->i915;
> >   	struct i2c_adapter *adapter = dvo->i2c_bus;
> >   	u8 out_buf[2];
> >   	struct i2c_msg msg = {
> > @@ -178,7 +178,7 @@ static bool tfp410_init(struct intel_dvo_device *dvo,
> >   		return false;
> >   	dvo->i2c_bus = adapter;
> > -	dvo->dev_priv = tfp;
> > +	dvo->i915 = tfp;
> >   	tfp->quiet = true;
> >   	if ((id = tfp410_getid(dvo, TFP410_VID_LO)) != TFP410_VID) {
> > @@ -299,11 +299,11 @@ static void tfp410_dump_regs(struct intel_dvo_device *dvo)
> >   static void tfp410_destroy(struct intel_dvo_device *dvo)
> >   {
> > -	struct tfp410_priv *tfp = dvo->dev_priv;
> > +	struct tfp410_priv *tfp = dvo->i915;
> >   	if (tfp) {
> >   		kfree(tfp);
> > -		dvo->dev_priv = NULL;
> > +		dvo->i915 = NULL;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c
> > index c86ca9f21532..2fa6747c7b6e 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c
> > @@ -495,27 +495,27 @@ static void __assign_ppgtt(struct i915_gem_context *ctx,
> >   }
> >   static struct i915_gem_context *
> > -i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags)
> > +i915_gem_create_context(struct drm_i915_private *i915, unsigned int flags)
> >   {
> >   	struct i915_gem_context *ctx;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE &&
> > -	    !HAS_EXECLISTS(dev_priv))
> > +	    !HAS_EXECLISTS(i915))
> >   		return ERR_PTR(-EINVAL);
> >   	/* Reap the most stale context */
> > -	contexts_free_first(dev_priv);
> > +	contexts_free_first(i915);
> > -	ctx = __create_context(dev_priv);
> > +	ctx = __create_context(i915);
> >   	if (IS_ERR(ctx))
> >   		return ctx;
> > -	if (HAS_FULL_PPGTT(dev_priv)) {
> > +	if (HAS_FULL_PPGTT(i915)) {
> >   		struct i915_ppgtt *ppgtt;
> > -		ppgtt = i915_ppgtt_create(dev_priv);
> > +		ppgtt = i915_ppgtt_create(i915);
> >   		if (IS_ERR(ppgtt)) {
> >   			DRM_DEBUG_DRIVER("PPGTT setup failed (%ld)\n",
> >   					 PTR_ERR(ppgtt));
> > @@ -530,7 +530,7 @@ i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags)
> >   	if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE) {
> >   		struct i915_timeline *timeline;
> > -		timeline = i915_timeline_create(dev_priv, NULL);
> > +		timeline = i915_timeline_create(i915, NULL);
> >   		if (IS_ERR(timeline)) {
> >   			context_close(ctx);
> >   			return ERR_CAST(timeline);
> > @@ -649,19 +649,19 @@ static bool needs_preempt_context(struct drm_i915_private *i915)
> >   	return HAS_EXECLISTS(i915);
> >   }
> > -int i915_gem_contexts_init(struct drm_i915_private *dev_priv)
> > +int i915_gem_contexts_init(struct drm_i915_private *i915)
> >   {
> >   	struct i915_gem_context *ctx;
> >   	/* Reassure ourselves we are only called once */
> > -	GEM_BUG_ON(dev_priv->kernel_context);
> > -	GEM_BUG_ON(dev_priv->preempt_context);
> > +	GEM_BUG_ON(i915->kernel_context);
> > +	GEM_BUG_ON(i915->preempt_context);
> > -	intel_engine_init_ctx_wa(dev_priv->engine[RCS0]);
> > -	init_contexts(dev_priv);
> > +	intel_engine_init_ctx_wa(i915->engine[RCS0]);
> > +	init_contexts(i915);
> >   	/* lowest priority; idle task */
> > -	ctx = i915_gem_context_create_kernel(dev_priv, I915_PRIORITY_MIN);
> > +	ctx = i915_gem_context_create_kernel(i915, I915_PRIORITY_MIN);
> >   	if (IS_ERR(ctx)) {
> >   		DRM_ERROR("Failed to create default global context\n");
> >   		return PTR_ERR(ctx);
> > @@ -675,31 +675,31 @@ int i915_gem_contexts_init(struct drm_i915_private *dev_priv)
> >   	 */
> >   	GEM_BUG_ON(ctx->hw_id);
> >   	GEM_BUG_ON(!atomic_read(&ctx->hw_id_pin_count));
> > -	dev_priv->kernel_context = ctx;
> > +	i915->kernel_context = ctx;
> >   	/* highest priority; preempting task */
> > -	if (needs_preempt_context(dev_priv)) {
> > -		ctx = i915_gem_context_create_kernel(dev_priv, INT_MAX);
> > +	if (needs_preempt_context(i915)) {
> > +		ctx = i915_gem_context_create_kernel(i915, INT_MAX);
> >   		if (!IS_ERR(ctx))
> > -			dev_priv->preempt_context = ctx;
> > +			i915->preempt_context = ctx;
> >   		else
> >   			DRM_ERROR("Failed to create preempt context; disabling preemption\n");
> >   	}
> >   	DRM_DEBUG_DRIVER("%s context support initialized\n",
> > -			 DRIVER_CAPS(dev_priv)->has_logical_contexts ?
> > +			 DRIVER_CAPS(i915)->has_logical_contexts ?
> >   			 "logical" : "fake");
> >   	return 0;
> >   }
> > -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv)
> > +void i915_gem_contexts_lost(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		intel_engine_lost_context(engine);
> >   }
> > @@ -2368,7 +2368,7 @@ int i915_gem_context_setparam_ioctl(struct drm_device *dev, void *data,
> >   int i915_gem_context_reset_stats_ioctl(struct drm_device *dev,
> >   				       void *data, struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_reset_stats *args = data;
> >   	struct i915_gem_context *ctx;
> >   	int ret;
> > @@ -2390,7 +2390,7 @@ int i915_gem_context_reset_stats_ioctl(struct drm_device *dev,
> >   	 */
> >   	if (capable(CAP_SYS_ADMIN))
> > -		args->reset_count = i915_reset_count(&dev_priv->gpu_error);
> > +		args->reset_count = i915_reset_count(&i915->gpu_error);
> >   	else
> >   		args->reset_count = 0;
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.h b/drivers/gpu/drm/i915/gem/i915_gem_context.h
> > index 630392c77e48..f55cb8eae87a 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.h
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.h
> > @@ -133,9 +133,9 @@ static inline bool i915_gem_context_is_kernel(struct i915_gem_context *ctx)
> >   }
> >   /* i915_gem_context.c */
> > -int __must_check i915_gem_contexts_init(struct drm_i915_private *dev_priv);
> > -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv);
> > -void i915_gem_contexts_fini(struct drm_i915_private *dev_priv);
> > +int __must_check i915_gem_contexts_init(struct drm_i915_private *i915);
> > +void i915_gem_contexts_lost(struct drm_i915_private *i915);
> > +void i915_gem_contexts_fini(struct drm_i915_private *i915);
> >   int i915_gem_context_open(struct drm_i915_private *i915,
> >   			  struct drm_file *file);
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> > index 528eea44dccf..83d64580a2f2 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> > @@ -2099,7 +2099,7 @@ static int eb_submit(struct i915_execbuffer *eb)
> >    * The engine index is returned.
> >    */
> >   static unsigned int
> > -gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv,
> > +gen8_dispatch_bsd_engine(struct drm_i915_private *i915,
> >   			 struct drm_file *file)
> >   {
> >   	struct drm_i915_file_private *file_priv = file->driver_priv;
> > @@ -2107,7 +2107,7 @@ gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv,
> >   	/* Check whether the file_priv has already selected one ring. */
> >   	if ((int)file_priv->bsd_engine < 0)
> >   		file_priv->bsd_engine = atomic_fetch_xor(1,
> > -			 &dev_priv->mm.bsd_engine_dispatch_index);
> > +			 &i915->mm.bsd_engine_dispatch_index);
> >   	return file_priv->bsd_engine;
> >   }
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c
> > index 36b76c6a0a9d..a37f0adac16e 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c
> > @@ -367,7 +367,7 @@ void
> >   i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj,
> >   				   unsigned int flush_domains)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct i915_vma *vma;
> >   	assert_object_held(obj);
> > @@ -377,7 +377,7 @@ i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj,
> >   	switch (obj->write_domain) {
> >   	case I915_GEM_DOMAIN_GTT:
> > -		i915_gem_flush_ggtt_writes(dev_priv);
> > +		i915_gem_flush_ggtt_writes(i915);
> >   		intel_fb_obj_flush(obj,
> >   				   fb_write_origin(obj, I915_GEM_DOMAIN_GTT));
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > index 19d9ecdb2894..eb4aae65f6e4 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > @@ -521,7 +521,7 @@ i915_gem_object_create_shmem(struct drm_i915_private *i915, u64 size)
> >   /* Allocate a new GEM object and fill it with the supplied data */
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_shmem_from_data(struct drm_i915_private *i915,
> >   				       const void *data, size_t size)
> >   {
> >   	struct drm_i915_gem_object *obj;
> > @@ -529,7 +529,7 @@ i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv,
> >   	size_t offset;
> >   	int err;
> > -	obj = i915_gem_object_create_shmem(dev_priv, round_up(size, PAGE_SIZE));
> > +	obj = i915_gem_object_create_shmem(i915, round_up(size, PAGE_SIZE));
> >   	if (IS_ERR(obj))
> >   		return obj;
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> > index de1fab2058ec..24e174f069f3 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> > @@ -24,48 +24,48 @@
> >    * for is a boon.
> >    */
> > -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv,
> > +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915,
> >   					 struct drm_mm_node *node, u64 size,
> >   					 unsigned alignment, u64 start, u64 end)
> >   {
> >   	int ret;
> > -	if (!drm_mm_initialized(&dev_priv->mm.stolen))
> > +	if (!drm_mm_initialized(&i915->mm.stolen))
> >   		return -ENODEV;
> >   	/* WaSkipStolenMemoryFirstPage:bdw+ */
> > -	if (INTEL_GEN(dev_priv) >= 8 && start < 4096)
> > +	if (INTEL_GEN(i915) >= 8 && start < 4096)
> >   		start = 4096;
> > -	mutex_lock(&dev_priv->mm.stolen_lock);
> > -	ret = drm_mm_insert_node_in_range(&dev_priv->mm.stolen, node,
> > +	mutex_lock(&i915->mm.stolen_lock);
> > +	ret = drm_mm_insert_node_in_range(&i915->mm.stolen, node,
> >   					  size, alignment, 0,
> >   					  start, end, DRM_MM_INSERT_BEST);
> > -	mutex_unlock(&dev_priv->mm.stolen_lock);
> > +	mutex_unlock(&i915->mm.stolen_lock);
> >   	return ret;
> >   }
> > -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv,
> > +int i915_gem_stolen_insert_node(struct drm_i915_private *i915,
> >   				struct drm_mm_node *node, u64 size,
> >   				unsigned alignment)
> >   {
> > -	return i915_gem_stolen_insert_node_in_range(dev_priv, node, size,
> > +	return i915_gem_stolen_insert_node_in_range(i915, node, size,
> >   						    alignment, 0, U64_MAX);
> >   }
> > -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv,
> > +void i915_gem_stolen_remove_node(struct drm_i915_private *i915,
> >   				 struct drm_mm_node *node)
> >   {
> > -	mutex_lock(&dev_priv->mm.stolen_lock);
> > +	mutex_lock(&i915->mm.stolen_lock);
> >   	drm_mm_remove_node(node);
> > -	mutex_unlock(&dev_priv->mm.stolen_lock);
> > +	mutex_unlock(&i915->mm.stolen_lock);
> >   }
> > -static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> > +static int i915_adjust_stolen(struct drm_i915_private *i915,
> >   			      struct resource *dsm)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	struct resource *r;
> >   	if (dsm->start == 0 || dsm->end <= dsm->start)
> > @@ -77,14 +77,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> >   	 */
> >   	/* Make sure we don't clobber the GTT if it's within stolen memory */
> > -	if (INTEL_GEN(dev_priv) <= 4 &&
> > -	    !IS_G33(dev_priv) && !IS_PINEVIEW(dev_priv) && !IS_G4X(dev_priv)) {
> > +	if (INTEL_GEN(i915) <= 4 &&
> > +	    !IS_G33(i915) && !IS_PINEVIEW(i915) && !IS_G4X(i915)) {
> >   		struct resource stolen[2] = {*dsm, *dsm};
> >   		struct resource ggtt_res;
> >   		resource_size_t ggtt_start;
> >   		ggtt_start = I915_READ(PGTBL_CTL);
> > -		if (IS_GEN(dev_priv, 4))
> > +		if (IS_GEN(i915, 4))
> >   			ggtt_start = (ggtt_start & PGTBL_ADDRESS_LO_MASK) |
> >   				     (ggtt_start & PGTBL_ADDRESS_HI_MASK) << 28;
> >   		else
> > @@ -118,7 +118,7 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> >   	 * kernel. So if the region is already marked as busy, something
> >   	 * is seriously wrong.
> >   	 */
> > -	r = devm_request_mem_region(dev_priv->drm.dev, dsm->start,
> > +	r = devm_request_mem_region(i915->drm.dev, dsm->start,
> >   				    resource_size(dsm),
> >   				    "Graphics Stolen Memory");
> >   	if (r == NULL) {
> > @@ -131,14 +131,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> >   		 * reservation starting from 1 instead of 0.
> >   		 * There's also BIOS with off-by-one on the other end.
> >   		 */
> > -		r = devm_request_mem_region(dev_priv->drm.dev, dsm->start + 1,
> > +		r = devm_request_mem_region(i915->drm.dev, dsm->start + 1,
> >   					    resource_size(dsm) - 2,
> >   					    "Graphics Stolen Memory");
> >   		/*
> >   		 * GEN3 firmware likes to smash pci bridges into the stolen
> >   		 * range. Apparently this works.
> >   		 */
> > -		if (r == NULL && !IS_GEN(dev_priv, 3)) {
> > +		if (r == NULL && !IS_GEN(i915, 3)) {
> >   			DRM_ERROR("conflict detected with stolen region: %pR\n",
> >   				  dsm);
> > @@ -149,25 +149,25 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv)
> > +void i915_gem_cleanup_stolen(struct drm_i915_private *i915)
> >   {
> > -	if (!drm_mm_initialized(&dev_priv->mm.stolen))
> > +	if (!drm_mm_initialized(&i915->mm.stolen))
> >   		return;
> > -	drm_mm_takedown(&dev_priv->mm.stolen);
> > +	drm_mm_takedown(&i915->mm.stolen);
> >   }
> > -static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void g4x_get_stolen_reserved(struct drm_i915_private *i915,
> >   				    resource_size_t *base,
> >   				    resource_size_t *size)
> >   {
> > -	u32 reg_val = I915_READ(IS_GM45(dev_priv) ?
> > +	u32 reg_val = I915_READ(IS_GM45(i915) ?
> >   				CTG_STOLEN_RESERVED :
> >   				ELK_STOLEN_RESERVED);
> > -	resource_size_t stolen_top = dev_priv->dsm.end + 1;
> > +	resource_size_t stolen_top = i915->dsm.end + 1;
> >   	DRM_DEBUG_DRIVER("%s_STOLEN_RESERVED = %08x\n",
> > -			 IS_GM45(dev_priv) ? "CTG" : "ELK", reg_val);
> > +			 IS_GM45(i915) ? "CTG" : "ELK", reg_val);
> >   	if ((reg_val & G4X_STOLEN_RESERVED_ENABLE) == 0)
> >   		return;
> > @@ -176,7 +176,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	 * Whether ILK really reuses the ELK register for this is unclear.
> >   	 * Let's see if we catch anyone with this supposedly enabled on ILK.
> >   	 */
> > -	WARN(IS_GEN(dev_priv, 5), "ILK stolen reserved found? 0x%08x\n",
> > +	WARN(IS_GEN(i915, 5), "ILK stolen reserved found? 0x%08x\n",
> >   	     reg_val);
> >   	if (!(reg_val & G4X_STOLEN_RESERVED_ADDR2_MASK))
> > @@ -188,7 +188,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	*size = stolen_top - *base;
> >   }
> > -static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void gen6_get_stolen_reserved(struct drm_i915_private *i915,
> >   				     resource_size_t *base,
> >   				     resource_size_t *size)
> >   {
> > @@ -220,12 +220,12 @@ static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void vlv_get_stolen_reserved(struct drm_i915_private *i915,
> >   				    resource_size_t *base,
> >   				    resource_size_t *size)
> >   {
> >   	u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED);
> > -	resource_size_t stolen_top = dev_priv->dsm.end + 1;
> > +	resource_size_t stolen_top = i915->dsm.end + 1;
> >   	DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val);
> > @@ -248,7 +248,7 @@ static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	*base = stolen_top - *size;
> >   }
> > -static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void gen7_get_stolen_reserved(struct drm_i915_private *i915,
> >   				     resource_size_t *base,
> >   				     resource_size_t *size)
> >   {
> > @@ -274,7 +274,7 @@ static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void chv_get_stolen_reserved(struct drm_i915_private *i915,
> >   				    resource_size_t *base,
> >   				    resource_size_t *size)
> >   {
> > @@ -306,12 +306,12 @@ static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void bdw_get_stolen_reserved(struct drm_i915_private *dev_priv,
> > +static void bdw_get_stolen_reserved(struct drm_i915_private *i915,
> >   				    resource_size_t *base,
> >   				    resource_size_t *size)
> >   {
> >   	u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED);
> > -	resource_size_t stolen_top = dev_priv->dsm.end + 1;
> > +	resource_size_t stolen_top = i915->dsm.end + 1;
> >   	DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val);
> > @@ -354,19 +354,19 @@ static void icl_get_stolen_reserved(struct drm_i915_private *i915,
> >   	}
> >   }
> > -int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> > +int i915_gem_init_stolen(struct drm_i915_private *i915)
> >   {
> >   	resource_size_t reserved_base, stolen_top;
> >   	resource_size_t reserved_total, reserved_size;
> > -	mutex_init(&dev_priv->mm.stolen_lock);
> > +	mutex_init(&i915->mm.stolen_lock);
> > -	if (intel_vgpu_active(dev_priv)) {
> > +	if (intel_vgpu_active(i915)) {
> >   		DRM_INFO("iGVT-g active, disabling use of stolen memory\n");
> >   		return 0;
> >   	}
> > -	if (intel_vtd_active() && INTEL_GEN(dev_priv) < 8) {
> > +	if (intel_vtd_active() && INTEL_GEN(i915) < 8) {
> >   		DRM_INFO("DMAR active, disabling use of stolen memory\n");
> >   		return 0;
> >   	}
> > @@ -374,55 +374,55 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> >   	if (resource_size(&intel_graphics_stolen_res) == 0)
> >   		return 0;
> > -	dev_priv->dsm = intel_graphics_stolen_res;
> > +	i915->dsm = intel_graphics_stolen_res;
> > -	if (i915_adjust_stolen(dev_priv, &dev_priv->dsm))
> > +	if (i915_adjust_stolen(i915, &i915->dsm))
> >   		return 0;
> > -	GEM_BUG_ON(dev_priv->dsm.start == 0);
> > -	GEM_BUG_ON(dev_priv->dsm.end <= dev_priv->dsm.start);
> > +	GEM_BUG_ON(i915->dsm.start == 0);
> > +	GEM_BUG_ON(i915->dsm.end <= i915->dsm.start);
> > -	stolen_top = dev_priv->dsm.end + 1;
> > +	stolen_top = i915->dsm.end + 1;
> >   	reserved_base = stolen_top;
> >   	reserved_size = 0;
> > -	switch (INTEL_GEN(dev_priv)) {
> > +	switch (INTEL_GEN(i915)) {
> >   	case 2:
> >   	case 3:
> >   		break;
> >   	case 4:
> > -		if (!IS_G4X(dev_priv))
> > +		if (!IS_G4X(i915))
> >   			break;
> >   		/* fall through */
> >   	case 5:
> > -		g4x_get_stolen_reserved(dev_priv,
> > +		g4x_get_stolen_reserved(i915,
> >   					&reserved_base, &reserved_size);
> >   		break;
> >   	case 6:
> > -		gen6_get_stolen_reserved(dev_priv,
> > +		gen6_get_stolen_reserved(i915,
> >   					 &reserved_base, &reserved_size);
> >   		break;
> >   	case 7:
> > -		if (IS_VALLEYVIEW(dev_priv))
> > -			vlv_get_stolen_reserved(dev_priv,
> > +		if (IS_VALLEYVIEW(i915))
> > +			vlv_get_stolen_reserved(i915,
> >   						&reserved_base, &reserved_size);
> >   		else
> > -			gen7_get_stolen_reserved(dev_priv,
> > +			gen7_get_stolen_reserved(i915,
> >   						 &reserved_base, &reserved_size);
> >   		break;
> >   	case 8:
> >   	case 9:
> >   	case 10:
> > -		if (IS_LP(dev_priv))
> > -			chv_get_stolen_reserved(dev_priv,
> > +		if (IS_LP(i915))
> > +			chv_get_stolen_reserved(i915,
> >   						&reserved_base, &reserved_size);
> >   		else
> > -			bdw_get_stolen_reserved(dev_priv,
> > +			bdw_get_stolen_reserved(i915,
> >   						&reserved_base, &reserved_size);
> >   		break;
> >   	case 11:
> >   	default:
> > -		icl_get_stolen_reserved(dev_priv, &reserved_base,
> > +		icl_get_stolen_reserved(i915, &reserved_base,
> >   					&reserved_size);
> >   		break;
> >   	}
> > @@ -439,12 +439,12 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> >   		reserved_size = 0;
> >   	}
> > -	dev_priv->dsm_reserved =
> > +	i915->dsm_reserved =
> >   		(struct resource) DEFINE_RES_MEM(reserved_base, reserved_size);
> > -	if (!resource_contains(&dev_priv->dsm, &dev_priv->dsm_reserved)) {
> > +	if (!resource_contains(&i915->dsm, &i915->dsm_reserved)) {
> >   		DRM_ERROR("Stolen reserved area %pR outside stolen memory %pR\n",
> > -			  &dev_priv->dsm_reserved, &dev_priv->dsm);
> > +			  &i915->dsm_reserved, &i915->dsm);
> >   		return 0;
> >   	}
> > @@ -453,14 +453,14 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv)
> >   	reserved_total = stolen_top - reserved_base;
> >   	DRM_DEBUG_DRIVER("Memory reserved for graphics device: %lluK, usable: %lluK\n",
> > -			 (u64)resource_size(&dev_priv->dsm) >> 10,
> > -			 ((u64)resource_size(&dev_priv->dsm) - reserved_total) >> 10);
> > +			 (u64)resource_size(&i915->dsm) >> 10,
> > +			 ((u64)resource_size(&i915->dsm) - reserved_total) >> 10);
> > -	dev_priv->stolen_usable_size =
> > -		resource_size(&dev_priv->dsm) - reserved_total;
> > +	i915->stolen_usable_size =
> > +		resource_size(&i915->dsm) - reserved_total;
> >   	/* Basic memrange allocator for stolen space. */
> > -	drm_mm_init(&dev_priv->mm.stolen, 0, dev_priv->stolen_usable_size);
> > +	drm_mm_init(&i915->mm.stolen, 0, i915->stolen_usable_size);
> >   	return 0;
> >   }
> > @@ -469,11 +469,11 @@ static struct sg_table *
> >   i915_pages_create_for_stolen(struct drm_device *dev,
> >   			     resource_size_t offset, resource_size_t size)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct sg_table *st;
> >   	struct scatterlist *sg;
> > -	GEM_BUG_ON(range_overflows(offset, size, resource_size(&dev_priv->dsm)));
> > +	GEM_BUG_ON(range_overflows(offset, size, resource_size(&i915->dsm)));
> >   	/* We hide that we have no struct page backing our stolen object
> >   	 * by wrapping the contiguous physical allocation with a fake
> > @@ -493,7 +493,7 @@ i915_pages_create_for_stolen(struct drm_device *dev,
> >   	sg->offset = 0;
> >   	sg->length = size;
> > -	sg_dma_address(sg) = (dma_addr_t)dev_priv->dsm.start + offset;
> > +	sg_dma_address(sg) = (dma_addr_t)i915->dsm.start + offset;
> >   	sg_dma_len(sg) = size;
> >   	return st;
> > @@ -524,14 +524,14 @@ static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
> >   static void
> >   i915_gem_object_release_stolen(struct drm_i915_gem_object *obj)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct drm_mm_node *stolen = fetch_and_zero(&obj->stolen);
> >   	GEM_BUG_ON(!stolen);
> >   	__i915_gem_object_unpin_pages(obj);
> > -	i915_gem_stolen_remove_node(dev_priv, stolen);
> > +	i915_gem_stolen_remove_node(i915, stolen);
> >   	kfree(stolen);
> >   }
> > @@ -542,7 +542,7 @@ static const struct drm_i915_gem_object_ops i915_gem_object_stolen_ops = {
> >   };
> >   static struct drm_i915_gem_object *
> > -_i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> > +_i915_gem_object_create_stolen(struct drm_i915_private *i915,
> >   			       struct drm_mm_node *stolen)
> >   {
> >   	struct drm_i915_gem_object *obj;
> > @@ -552,12 +552,12 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> >   	if (obj == NULL)
> >   		return NULL;
> > -	drm_gem_private_object_init(&dev_priv->drm, &obj->base, stolen->size);
> > +	drm_gem_private_object_init(&i915->drm, &obj->base, stolen->size);
> >   	i915_gem_object_init(obj, &i915_gem_object_stolen_ops);
> >   	obj->stolen = stolen;
> >   	obj->read_domains = I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT;
> > -	cache_level = HAS_LLC(dev_priv) ? I915_CACHE_LLC : I915_CACHE_NONE;
> > +	cache_level = HAS_LLC(i915) ? I915_CACHE_LLC : I915_CACHE_NONE;
> >   	i915_gem_object_set_cache_coherency(obj, cache_level);
> >   	if (i915_gem_object_pin_pages(obj))
> > @@ -571,14 +571,14 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> >   }
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_stolen(struct drm_i915_private *i915,
> >   			      resource_size_t size)
> >   {
> >   	struct drm_i915_gem_object *obj;
> >   	struct drm_mm_node *stolen;
> >   	int ret;
> > -	if (!drm_mm_initialized(&dev_priv->mm.stolen))
> > +	if (!drm_mm_initialized(&i915->mm.stolen))
> >   		return NULL;
> >   	if (size == 0)
> > @@ -588,37 +588,37 @@ i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> >   	if (!stolen)
> >   		return NULL;
> > -	ret = i915_gem_stolen_insert_node(dev_priv, stolen, size, 4096);
> > +	ret = i915_gem_stolen_insert_node(i915, stolen, size, 4096);
> >   	if (ret) {
> >   		kfree(stolen);
> >   		return NULL;
> >   	}
> > -	obj = _i915_gem_object_create_stolen(dev_priv, stolen);
> > +	obj = _i915_gem_object_create_stolen(i915, stolen);
> >   	if (obj)
> >   		return obj;
> > -	i915_gem_stolen_remove_node(dev_priv, stolen);
> > +	i915_gem_stolen_remove_node(i915, stolen);
> >   	kfree(stolen);
> >   	return NULL;
> >   }
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915,
> >   					       resource_size_t stolen_offset,
> >   					       resource_size_t gtt_offset,
> >   					       resource_size_t size)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	struct drm_i915_gem_object *obj;
> >   	struct drm_mm_node *stolen;
> >   	struct i915_vma *vma;
> >   	int ret;
> > -	if (!drm_mm_initialized(&dev_priv->mm.stolen))
> > +	if (!drm_mm_initialized(&i915->mm.stolen))
> >   		return NULL;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	DRM_DEBUG_DRIVER("creating preallocated stolen object: stolen_offset=%pa, gtt_offset=%pa, size=%pa\n",
> >   			 &stolen_offset, &gtt_offset, &size);
> > @@ -635,19 +635,19 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv
> >   	stolen->start = stolen_offset;
> >   	stolen->size = size;
> > -	mutex_lock(&dev_priv->mm.stolen_lock);
> > -	ret = drm_mm_reserve_node(&dev_priv->mm.stolen, stolen);
> > -	mutex_unlock(&dev_priv->mm.stolen_lock);
> > +	mutex_lock(&i915->mm.stolen_lock);
> > +	ret = drm_mm_reserve_node(&i915->mm.stolen, stolen);
> > +	mutex_unlock(&i915->mm.stolen_lock);
> >   	if (ret) {
> >   		DRM_DEBUG_DRIVER("failed to allocate stolen space\n");
> >   		kfree(stolen);
> >   		return NULL;
> >   	}
> > -	obj = _i915_gem_object_create_stolen(dev_priv, stolen);
> > +	obj = _i915_gem_object_create_stolen(i915, stolen);
> >   	if (obj == NULL) {
> >   		DRM_DEBUG_DRIVER("failed to allocate stolen object\n");
> > -		i915_gem_stolen_remove_node(dev_priv, stolen);
> > +		i915_gem_stolen_remove_node(i915, stolen);
> >   		kfree(stolen);
> >   		return NULL;
> >   	}
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> > index ca0c2f451742..46b523d4ed84 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c
> > @@ -398,7 +398,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
> >   			  struct drm_file *file)
> >   {
> >   	struct drm_i915_gem_get_tiling *args = data;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_object *obj;
> >   	int err = -ENOENT;
> > @@ -415,10 +415,10 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
> >   	switch (args->tiling_mode) {
> >   	case I915_TILING_X:
> > -		args->swizzle_mode = dev_priv->mm.bit_6_swizzle_x;
> > +		args->swizzle_mode = i915->mm.bit_6_swizzle_x;
> >   		break;
> >   	case I915_TILING_Y:
> > -		args->swizzle_mode = dev_priv->mm.bit_6_swizzle_y;
> > +		args->swizzle_mode = i915->mm.bit_6_swizzle_y;
> >   		break;
> >   	default:
> >   	case I915_TILING_NONE:
> > @@ -427,7 +427,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data,
> >   	}
> >   	/* Hide bit 17 from the user -- see comment in i915_gem_set_tiling */
> > -	if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> > +	if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> >   		args->phys_swizzle_mode = I915_BIT_6_SWIZZLE_UNKNOWN;
> >   	else
> >   		args->phys_swizzle_mode = args->swizzle_mode;
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > index 528b61678334..f22e10e8895a 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > @@ -327,12 +327,12 @@ i915_mmu_notifier_free(struct i915_mmu_notifier *mn,
> >   #endif
> >   static struct i915_mm_struct *
> > -__i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real)
> > +__i915_mm_struct_find(struct drm_i915_private *i915, struct mm_struct *real)
> >   {
> >   	struct i915_mm_struct *mm;
> > -	/* Protected by dev_priv->mm_lock */
> > -	hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real)
> > +	/* Protected by i915->mm_lock */
> > +	hash_for_each_possible(i915->mm_structs, mm, node, (unsigned long)real)
> >   		if (mm->mm == real)
> >   			return mm;
> > @@ -342,7 +342,7 @@ __i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real)
> >   static int
> >   i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct i915_mm_struct *mm;
> >   	int ret = 0;
> > @@ -356,8 +356,8 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> >   	 * struct_mutex, i.e. we need to schedule a worker to do the clean
> >   	 * up.
> >   	 */
> > -	mutex_lock(&dev_priv->mm_lock);
> > -	mm = __i915_mm_struct_find(dev_priv, current->mm);
> > +	mutex_lock(&i915->mm_lock);
> > +	mm = __i915_mm_struct_find(i915, current->mm);
> >   	if (mm == NULL) {
> >   		mm = kmalloc(sizeof(*mm), GFP_KERNEL);
> >   		if (mm == NULL) {
> > @@ -373,15 +373,15 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> >   		mm->mn = NULL;
> > -		/* Protected by dev_priv->mm_lock */
> > -		hash_add(dev_priv->mm_structs,
> > +		/* Protected by i915->mm_lock */
> > +		hash_add(i915->mm_structs,
> >   			 &mm->node, (unsigned long)mm->mm);
> >   	} else
> >   		kref_get(&mm->kref);
> >   	obj->userptr.mm = mm;
> >   out:
> > -	mutex_unlock(&dev_priv->mm_lock);
> > +	mutex_unlock(&i915->mm_lock);
> >   	return ret;
> >   }
> > @@ -399,7 +399,7 @@ __i915_mm_struct_free(struct kref *kref)
> >   {
> >   	struct i915_mm_struct *mm = container_of(kref, typeof(*mm), kref);
> > -	/* Protected by dev_priv->mm_lock */
> > +	/* Protected by i915->mm_lock */
> >   	hash_del(&mm->node);
> >   	mutex_unlock(&mm->i915->mm_lock);
> > @@ -741,13 +741,13 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> >   		       void *data,
> >   		       struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_userptr *args = data;
> >   	struct drm_i915_gem_object *obj;
> >   	int ret;
> >   	u32 handle;
> > -	if (!HAS_LLC(dev_priv) && !HAS_SNOOP(dev_priv)) {
> > +	if (!HAS_LLC(i915) && !HAS_SNOOP(i915)) {
> >   		/* We cannot support coherent userptr objects on hw without
> >   		 * LLC and broken snooping.
> >   		 */
> > @@ -774,7 +774,7 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> >   		 * On almost all of the older hw, we cannot tell the GPU that
> >   		 * a page is readonly.
> >   		 */
> > -		vm = dev_priv->kernel_context->vm;
> > +		vm = i915->kernel_context->vm;
> >   		if (!vm || !vm->has_read_only)
> >   			return -ENODEV;
> >   	}
> > @@ -812,22 +812,22 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> >   	return 0;
> >   }
> > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv)
> > +int i915_gem_init_userptr(struct drm_i915_private *i915)
> >   {
> > -	mutex_init(&dev_priv->mm_lock);
> > -	hash_init(dev_priv->mm_structs);
> > +	mutex_init(&i915->mm_lock);
> > +	hash_init(i915->mm_structs);
> > -	dev_priv->mm.userptr_wq =
> > +	i915->mm.userptr_wq =
> >   		alloc_workqueue("i915-userptr-acquire",
> >   				WQ_HIGHPRI | WQ_UNBOUND,
> >   				0);
> > -	if (!dev_priv->mm.userptr_wq)
> > +	if (!i915->mm.userptr_wq)
> >   		return -ENOMEM;
> >   	return 0;
> >   }
> > -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv)
> > +void i915_gem_cleanup_userptr(struct drm_i915_private *i915)
> >   {
> > -	destroy_workqueue(dev_priv->mm.userptr_wq);
> > +	destroy_workqueue(i915->mm.userptr_wq);
> >   }
> > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> > index 73e667b31cc4..78797699feb0 100644
> > --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> > @@ -1415,8 +1415,8 @@ static int igt_ppgtt_gemfs_huge(void *arg)
> >   static int igt_ppgtt_pin_update(void *arg)
> >   {
> >   	struct i915_gem_context *ctx = arg;
> > -	struct drm_i915_private *dev_priv = ctx->i915;
> > -	unsigned long supported = INTEL_INFO(dev_priv)->page_sizes;
> > +	struct drm_i915_private *i915 = ctx->i915;
> > +	unsigned long supported = INTEL_INFO(i915)->page_sizes;
> >   	struct i915_address_space *vm = ctx->vm;
> >   	struct drm_i915_gem_object *obj;
> >   	struct i915_vma *vma;
> > @@ -1443,7 +1443,7 @@ static int igt_ppgtt_pin_update(void *arg)
> >   	for_each_set_bit_from(first, &supported, last + 1) {
> >   		unsigned int page_size = BIT(first);
> > -		obj = i915_gem_object_create_internal(dev_priv, page_size);
> > +		obj = i915_gem_object_create_internal(i915, page_size);
> >   		if (IS_ERR(obj))
> >   			return PTR_ERR(obj);
> > @@ -1497,7 +1497,7 @@ static int igt_ppgtt_pin_update(void *arg)
> >   		i915_gem_object_put(obj);
> >   	}
> > -	obj = i915_gem_object_create_internal(dev_priv, PAGE_SIZE);
> > +	obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
> >   	if (IS_ERR(obj))
> >   		return PTR_ERR(obj);
> > @@ -1518,7 +1518,7 @@ static int igt_ppgtt_pin_update(void *arg)
> >   	 * land in the now stale 2M page.
> >   	 */
> > -	err = gpu_write(vma, ctx, dev_priv->engine[RCS0], 0, 0xdeadbeaf);
> > +	err = gpu_write(vma, ctx, i915->engine[RCS0], 0, 0xdeadbeaf);
> >   	if (err)
> >   		goto out_unpin;
> > @@ -1682,20 +1682,20 @@ int i915_gem_huge_page_mock_selftests(void)
> >   		SUBTEST(igt_mock_ppgtt_huge_fill),
> >   		SUBTEST(igt_mock_ppgtt_64K),
> >   	};
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	struct i915_ppgtt *ppgtt;
> >   	int err;
> > -	dev_priv = mock_gem_device();
> > -	if (!dev_priv)
> > +	i915 = mock_gem_device();
> > +	if (!i915)
> >   		return -ENOMEM;
> >   	/* Pretend to be a device which supports the 48b PPGTT */
> > -	mkwrite_device_info(dev_priv)->ppgtt_type = INTEL_PPGTT_FULL;
> > -	mkwrite_device_info(dev_priv)->ppgtt_size = 48;
> > +	mkwrite_device_info(i915)->ppgtt_type = INTEL_PPGTT_FULL;
> > +	mkwrite_device_info(i915)->ppgtt_size = 48;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	ppgtt = i915_ppgtt_create(dev_priv);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	ppgtt = i915_ppgtt_create(i915);
> >   	if (IS_ERR(ppgtt)) {
> >   		err = PTR_ERR(ppgtt);
> >   		goto out_unlock;
> > @@ -1720,13 +1720,13 @@ int i915_gem_huge_page_mock_selftests(void)
> >   	i915_vm_put(&ppgtt->vm);
> >   out_unlock:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > -	drm_dev_put(&dev_priv->drm);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > +	drm_dev_put(&i915->drm);
> >   	return err;
> >   }
> > -int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> > +int i915_gem_huge_page_live_selftests(struct drm_i915_private *i915)
> >   {
> >   	static const struct i915_subtest tests[] = {
> >   		SUBTEST(igt_shrink_thp),
> > @@ -1741,22 +1741,22 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> >   	intel_wakeref_t wakeref;
> >   	int err;
> > -	if (!HAS_PPGTT(dev_priv)) {
> > +	if (!HAS_PPGTT(i915)) {
> >   		pr_info("PPGTT not supported, skipping live-selftests\n");
> >   		return 0;
> >   	}
> > -	if (i915_terminally_wedged(dev_priv))
> > +	if (i915_terminally_wedged(i915))
> >   		return 0;
> > -	file = mock_file(dev_priv);
> > +	file = mock_file(i915);
> >   	if (IS_ERR(file))
> >   		return PTR_ERR(file);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	ctx = live_context(dev_priv, file);
> > +	ctx = live_context(i915, file);
> >   	if (IS_ERR(ctx)) {
> >   		err = PTR_ERR(ctx);
> >   		goto out_unlock;
> > @@ -1768,10 +1768,10 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv)
> >   	err = i915_subtests(tests, ctx);
> >   out_unlock:
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	intel_runtime_pm_put(i915, wakeref);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	mock_file_free(dev_priv, file);
> > +	mock_file_free(i915, file);
> >   	return err;
> >   }
> > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > index 74b0e5871c4b..2fdcaa1667bf 100644
> > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > @@ -1748,7 +1748,7 @@ int i915_gem_context_mock_selftests(void)
> >   	return err;
> >   }
> > -int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv)
> > +int i915_gem_context_live_selftests(struct drm_i915_private *i915)
> >   {
> >   	static const struct i915_subtest tests[] = {
> >   		SUBTEST(live_nop_switch),
> > @@ -1759,8 +1759,8 @@ int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv)
> >   		SUBTEST(igt_vm_isolation),
> >   	};
> > -	if (i915_terminally_wedged(dev_priv))
> > +	if (i915_terminally_wedged(i915))
> >   		return 0;
> > -	return i915_subtests(tests, dev_priv);
> > +	return i915_subtests(tests, i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/gt/intel_engine.h b/drivers/gpu/drm/i915/gt/intel_engine.h
> > index 1439fa4093ac..74f9ad6648eb 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_engine.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_engine.h
> > @@ -465,7 +465,7 @@ static inline void intel_engine_reset(struct intel_engine_cs *engine,
> >   }
> >   bool intel_engine_is_idle(struct intel_engine_cs *engine);
> > -bool intel_engines_are_idle(struct drm_i915_private *dev_priv);
> > +bool intel_engines_are_idle(struct drm_i915_private *i915);
> >   void intel_engine_lost_context(struct intel_engine_cs *engine);
> > diff --git a/drivers/gpu/drm/i915/gt/intel_engine_cs.c b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> > index c0d986db5a75..a048b8743ce6 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_engine_cs.c
> > @@ -160,7 +160,7 @@ static const struct engine_info intel_engines[] = {
> >   /**
> >    * intel_engine_context_size() - return the size of the context for an engine
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    * @class: engine class
> >    *
> >    * Each engine class may require a different amount of space for a context
> > @@ -172,7 +172,7 @@ static const struct engine_info intel_engines[] = {
> >    * in LRC mode, but does not include the "shared data page" used with
> >    * GuC submission. The caller should account for this if using the GuC.
> >    */
> > -u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> > +u32 intel_engine_context_size(struct drm_i915_private *i915, u8 class)
> >   {
> >   	u32 cxt_size;
> > @@ -180,9 +180,9 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> >   	switch (class) {
> >   	case RENDER_CLASS:
> > -		switch (INTEL_GEN(dev_priv)) {
> > +		switch (INTEL_GEN(i915)) {
> >   		default:
> > -			MISSING_CASE(INTEL_GEN(dev_priv));
> > +			MISSING_CASE(INTEL_GEN(i915));
> >   			return DEFAULT_LR_CONTEXT_RENDER_SIZE;
> >   		case 11:
> >   			return GEN11_LR_CONTEXT_RENDER_SIZE;
> > @@ -193,7 +193,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> >   		case 8:
> >   			return GEN8_LR_CONTEXT_RENDER_SIZE;
> >   		case 7:
> > -			if (IS_HASWELL(dev_priv))
> > +			if (IS_HASWELL(i915))
> >   				return HSW_CXT_TOTAL_SIZE;
> >   			cxt_size = I915_READ(GEN7_CXT_SIZE);
> > @@ -217,7 +217,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> >   			 */
> >   			cxt_size = I915_READ(CXT_SIZE) + 1;
> >   			DRM_DEBUG_DRIVER("gen%d CXT_SIZE = %d bytes [0x%08x]\n",
> > -					 INTEL_GEN(dev_priv),
> > +					 INTEL_GEN(i915),
> >   					 cxt_size * 64,
> >   					 cxt_size - 1);
> >   			return round_up(cxt_size * 64, PAGE_SIZE);
> > @@ -234,7 +234,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class)
> >   	case VIDEO_DECODE_CLASS:
> >   	case VIDEO_ENHANCEMENT_CLASS:
> >   	case COPY_ENGINE_CLASS:
> > -		if (INTEL_GEN(dev_priv) < 8)
> > +		if (INTEL_GEN(i915) < 8)
> >   			return 0;
> >   		return GEN8_LR_CONTEXT_OTHER_SIZE;
> >   	}
> > @@ -284,7 +284,7 @@ static void intel_engine_sanitize_mmio(struct intel_engine_cs *engine)
> >   }
> >   static int
> > -intel_engine_setup(struct drm_i915_private *dev_priv,
> > +intel_engine_setup(struct drm_i915_private *i915,
> >   		   enum intel_engine_id id)
> >   {
> >   	const struct engine_info *info = &intel_engines[id];
> > @@ -301,10 +301,10 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> >   	if (GEM_DEBUG_WARN_ON(info->instance > MAX_ENGINE_INSTANCE))
> >   		return -EINVAL;
> > -	if (GEM_DEBUG_WARN_ON(dev_priv->engine_class[info->class][info->instance]))
> > +	if (GEM_DEBUG_WARN_ON(i915->engine_class[info->class][info->instance]))
> >   		return -EINVAL;
> > -	GEM_BUG_ON(dev_priv->engine[id]);
> > +	GEM_BUG_ON(i915->engine[id]);
> >   	engine = kzalloc(sizeof(*engine), GFP_KERNEL);
> >   	if (!engine)
> >   		return -ENOMEM;
> > @@ -313,11 +313,11 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> >   	engine->id = id;
> >   	engine->mask = BIT(id);
> > -	engine->i915 = dev_priv;
> > -	engine->uncore = &dev_priv->uncore;
> > +	engine->i915 = i915;
> > +	engine->uncore = &i915->uncore;
> >   	__sprint_engine_name(engine->name, info);
> >   	engine->hw_id = engine->guc_id = info->hw_id;
> > -	engine->mmio_base = __engine_mmio_base(dev_priv, info->mmio_bases);
> > +	engine->mmio_base = __engine_mmio_base(i915, info->mmio_bases);
> >   	engine->class = info->class;
> >   	engine->instance = info->instance;
> > @@ -329,12 +329,12 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> >   	engine->uabi_class = intel_engine_classes[info->class].uabi_class;
> > -	engine->context_size = intel_engine_context_size(dev_priv,
> > +	engine->context_size = intel_engine_context_size(i915,
> >   							 engine->class);
> >   	if (WARN_ON(engine->context_size > BIT(20)))
> >   		engine->context_size = 0;
> >   	if (engine->context_size)
> > -		DRIVER_CAPS(dev_priv)->has_logical_contexts = true;
> > +		DRIVER_CAPS(i915)->has_logical_contexts = true;
> >   	/* Nothing to do here, execute in order of dependencies */
> >   	engine->schedule = NULL;
> > @@ -346,8 +346,8 @@ intel_engine_setup(struct drm_i915_private *dev_priv,
> >   	/* Scrub mmio state on takeover */
> >   	intel_engine_sanitize_mmio(engine);
> > -	dev_priv->engine_class[info->class][info->instance] = engine;
> > -	dev_priv->engine[id] = engine;
> > +	i915->engine_class[info->class][info->instance] = engine;
> > +	i915->engine[id] = engine;
> >   	return 0;
> >   }
> > @@ -953,17 +953,17 @@ const char *i915_cache_level_str(struct drm_i915_private *i915, int type)
> >   	}
> >   }
> > -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv)
> > +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915)
> >   {
> > -	const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	u32 mcr_s_ss_select;
> >   	u32 slice = fls(sseu->slice_mask);
> >   	u32 subslice = fls(sseu->subslice_mask[slice]);
> > -	if (IS_GEN(dev_priv, 10))
> > +	if (IS_GEN(i915, 10))
> >   		mcr_s_ss_select = GEN8_MCR_SLICE(slice) |
> >   				  GEN8_MCR_SUBSLICE(subslice);
> > -	else if (INTEL_GEN(dev_priv) >= 11)
> > +	else if (INTEL_GEN(i915) >= 11)
> >   		mcr_s_ss_select = GEN11_MCR_SLICE(slice) |
> >   				  GEN11_MCR_SUBSLICE(subslice);
> >   	else
> > @@ -1095,7 +1095,7 @@ void intel_engine_get_instdone(struct intel_engine_cs *engine,
> >   static bool ring_is_idle(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	intel_wakeref_t wakeref;
> >   	bool idle = true;
> > @@ -1103,7 +1103,7 @@ static bool ring_is_idle(struct intel_engine_cs *engine)
> >   		return true;
> >   	/* If the whole device is asleep, the engine must be idle */
> > -	wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> > +	wakeref = intel_runtime_pm_get_if_in_use(i915);
> >   	if (!wakeref)
> >   		return true;
> > @@ -1113,11 +1113,11 @@ static bool ring_is_idle(struct intel_engine_cs *engine)
> >   		idle = false;
> >   	/* No bit for gen2, so assume the CS parser is idle */
> > -	if (INTEL_GEN(dev_priv) > 2 &&
> > +	if (INTEL_GEN(i915) > 2 &&
> >   	    !(ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE))
> >   		idle = false;
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return idle;
> >   }
> > @@ -1321,12 +1321,12 @@ static void hexdump(struct drm_printer *m, const void *buf, size_t len)
> >   static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> >   					 struct drm_printer *m)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	const struct intel_engine_execlists * const execlists =
> >   		&engine->execlists;
> >   	u64 addr;
> > -	if (engine->id == RCS0 && IS_GEN_RANGE(dev_priv, 4, 7))
> > +	if (engine->id == RCS0 && IS_GEN_RANGE(i915, 4, 7))
> >   		drm_printf(m, "\tCCID: 0x%08x\n", ENGINE_READ(engine, CCID));
> >   	drm_printf(m, "\tRING_START: 0x%08x\n",
> >   		   ENGINE_READ(engine, RING_START));
> > @@ -1343,7 +1343,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> >   			   ENGINE_READ(engine, RING_MI_MODE) & (MODE_IDLE) ? " [idle]" : "");
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 6) {
> > +	if (INTEL_GEN(i915) >= 6) {
> >   		drm_printf(m, "\tRING_IMR: %08x\n",
> >   			   ENGINE_READ(engine, RING_IMR));
> >   	}
> > @@ -1354,15 +1354,15 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> >   	addr = intel_engine_get_last_batch_head(engine);
> >   	drm_printf(m, "\tBBADDR: 0x%08x_%08x\n",
> >   		   upper_32_bits(addr), lower_32_bits(addr));
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		addr = ENGINE_READ64(engine, RING_DMA_FADD, RING_DMA_FADD_UDW);
> > -	else if (INTEL_GEN(dev_priv) >= 4)
> > +	else if (INTEL_GEN(i915) >= 4)
> >   		addr = ENGINE_READ(engine, RING_DMA_FADD);
> >   	else
> >   		addr = ENGINE_READ(engine, DMA_FADD_I8XX);
> >   	drm_printf(m, "\tDMA_FADDR: 0x%08x_%08x\n",
> >   		   upper_32_bits(addr), lower_32_bits(addr));
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		drm_printf(m, "\tIPEIR: 0x%08x\n",
> >   			   ENGINE_READ(engine, RING_IPEIR));
> >   		drm_printf(m, "\tIPEHR: 0x%08x\n",
> > @@ -1372,7 +1372,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> >   		drm_printf(m, "\tIPEHR: 0x%08x\n", ENGINE_READ(engine, IPEHR));
> >   	}
> > -	if (HAS_EXECLISTS(dev_priv)) {
> > +	if (HAS_EXECLISTS(i915)) {
> >   		const u32 *hws =
> >   			&engine->status_page.addr[I915_HWS_CSB_BUF0_INDEX];
> >   		const u8 num_entries = execlists->csb_size;
> > @@ -1426,7 +1426,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine,
> >   		}
> >   		drm_printf(m, "\t\tHW active? 0x%x\n", execlists->active);
> >   		rcu_read_unlock();
> > -	} else if (INTEL_GEN(dev_priv) > 6) {
> > +	} else if (INTEL_GEN(i915) > 6) {
> >   		drm_printf(m, "\tPP_DIR_BASE: 0x%08x\n",
> >   			   ENGINE_READ(engine, RING_PP_DIR_BASE));
> >   		drm_printf(m, "\tPP_DIR_BASE_READ: 0x%08x\n",
> > diff --git a/drivers/gpu/drm/i915/gt/intel_hangcheck.c b/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> > index 174bb0a60309..948423c5f67d 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_hangcheck.c
> > @@ -50,7 +50,7 @@ static bool instdone_unchanged(u32 current_instdone, u32 *old_instdone)
> >   static bool subunits_stuck(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	struct intel_instdone instdone;
> >   	struct intel_instdone *accu_instdone = &engine->hangcheck.instdone;
> >   	bool stuck;
> > @@ -72,7 +72,7 @@ static bool subunits_stuck(struct intel_engine_cs *engine)
> >   	stuck &= instdone_unchanged(instdone.slice_common,
> >   				    &accu_instdone->slice_common);
> > -	for_each_instdone_slice_subslice(dev_priv, slice, subslice) {
> > +	for_each_instdone_slice_subslice(i915, slice, subslice) {
> >   		stuck &= instdone_unchanged(instdone.sampler[slice][subslice],
> >   					    &accu_instdone->sampler[slice][subslice]);
> >   		stuck &= instdone_unchanged(instdone.row[slice][subslice],
> > @@ -103,7 +103,7 @@ head_stuck(struct intel_engine_cs *engine, u64 acthd)
> >   static enum intel_engine_hangcheck_action
> >   engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	enum intel_engine_hangcheck_action ha;
> >   	u32 tmp;
> > @@ -111,7 +111,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> >   	if (ha != ENGINE_DEAD)
> >   		return ha;
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		return ENGINE_DEAD;
> >   	/* Is the chip hanging on a WAIT_FOR_EVENT?
> > @@ -121,7 +121,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd)
> >   	 */
> >   	tmp = ENGINE_READ(engine, RING_CTL);
> >   	if (tmp & RING_WAIT) {
> > -		i915_handle_error(dev_priv, engine->mask, 0,
> > +		i915_handle_error(i915, engine->mask, 0,
> >   				  "stuck wait on %s", engine->name);
> >   		ENGINE_WRITE(engine, RING_CTL, tmp);
> >   		return ENGINE_WAIT_KICK;
> > @@ -256,8 +256,8 @@ static void hangcheck_declare_hang(struct drm_i915_private *i915,
> >    */
> >   static void i915_hangcheck_elapsed(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv),
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915),
> >   			     gpu_error.hangcheck_work.work);
> >   	intel_engine_mask_t hung = 0, stuck = 0, wedged = 0;
> >   	struct intel_engine_cs *engine;
> > @@ -267,13 +267,13 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> >   	if (!i915_modparams.enable_hangcheck)
> >   		return;
> > -	if (!READ_ONCE(dev_priv->gt.awake))
> > +	if (!READ_ONCE(i915->gt.awake))
> >   		return;
> > -	if (i915_terminally_wedged(dev_priv))
> > +	if (i915_terminally_wedged(i915))
> >   		return;
> > -	wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> > +	wakeref = intel_runtime_pm_get_if_in_use(i915);
> >   	if (!wakeref)
> >   		return;
> > @@ -281,9 +281,9 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> >   	 * periodically arm the mmio checker to see if we are triggering
> >   	 * any invalid access.
> >   	 */
> > -	intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore);
> > +	intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore);
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		struct hangcheck hc;
> >   		intel_engine_signal_breadcrumbs(engine);
> > @@ -305,7 +305,7 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> >   	if (GEM_SHOW_DEBUG() && (hung | stuck)) {
> >   		struct drm_printer p = drm_debug_printer("hangcheck");
> > -		for_each_engine(engine, dev_priv, id) {
> > +		for_each_engine(engine, i915, id) {
> >   			if (intel_engine_is_idle(engine))
> >   				continue;
> > @@ -314,20 +314,20 @@ static void i915_hangcheck_elapsed(struct work_struct *work)
> >   	}
> >   	if (wedged) {
> > -		dev_err(dev_priv->drm.dev,
> > +		dev_err(i915->drm.dev,
> >   			"GPU recovery timed out,"
> >   			" cancelling all in-flight rendering.\n");
> >   		GEM_TRACE_DUMP();
> > -		i915_gem_set_wedged(dev_priv);
> > +		i915_gem_set_wedged(i915);
> >   	}
> >   	if (hung)
> > -		hangcheck_declare_hang(dev_priv, hung, stuck);
> > +		hangcheck_declare_hang(i915, hung, stuck);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	/* Reset timer in case GPU hangs without another request being added */
> > -	i915_queue_hangcheck(dev_priv);
> > +	i915_queue_hangcheck(i915);
> >   }
> >   void intel_engine_init_hangcheck(struct intel_engine_cs *engine)
> > diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.c b/drivers/gpu/drm/i915/gt/intel_mocs.c
> > index 79df66022d3a..9d6f0893f528 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_mocs.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_mocs.c
> > @@ -239,7 +239,7 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = {
> >   /**
> >    * get_mocs_settings()
> > - * @dev_priv:	i915 device.
> > + * @i915:	i915 device.
> >    * @table:      Output table that will be made to point at appropriate
> >    *	      MOCS values for the device.
> >    *
> > @@ -249,33 +249,33 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = {
> >    *
> >    * Return: true if there are applicable MOCS settings for the device.
> >    */
> > -static bool get_mocs_settings(struct drm_i915_private *dev_priv,
> > +static bool get_mocs_settings(struct drm_i915_private *i915,
> >   			      struct drm_i915_mocs_table *table)
> >   {
> >   	bool result = false;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		table->size  = ARRAY_SIZE(icelake_mocs_table);
> >   		table->table = icelake_mocs_table;
> >   		table->n_entries = GEN11_NUM_MOCS_ENTRIES;
> >   		result = true;
> > -	} else if (IS_GEN9_BC(dev_priv) || IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_GEN9_BC(i915) || IS_CANNONLAKE(i915)) {
> >   		table->size  = ARRAY_SIZE(skylake_mocs_table);
> >   		table->n_entries = GEN9_NUM_MOCS_ENTRIES;
> >   		table->table = skylake_mocs_table;
> >   		result = true;
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > +	} else if (IS_GEN9_LP(i915)) {
> >   		table->size  = ARRAY_SIZE(broxton_mocs_table);
> >   		table->n_entries = GEN9_NUM_MOCS_ENTRIES;
> >   		table->table = broxton_mocs_table;
> >   		result = true;
> >   	} else {
> > -		WARN_ONCE(INTEL_GEN(dev_priv) >= 9,
> > +		WARN_ONCE(INTEL_GEN(i915) >= 9,
> >   			  "Platform that should have a MOCS table does not.\n");
> >   	}
> >   	/* WaDisableSkipCaching:skl,bxt,kbl,glk */
> > -	if (IS_GEN(dev_priv, 9)) {
> > +	if (IS_GEN(i915, 9)) {
> >   		int i;
> >   		for (i = 0; i < table->size; i++)
> > @@ -330,12 +330,12 @@ static u32 get_entry_control(const struct drm_i915_mocs_table *table,
> >    */
> >   void intel_mocs_init_engine(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	struct drm_i915_mocs_table table;
> >   	unsigned int index;
> >   	u32 unused_value;
> > -	if (!get_mocs_settings(dev_priv, &table))
> > +	if (!get_mocs_settings(i915, &table))
> >   		return;
> >   	/* Set unused values to PTE */
> > @@ -482,7 +482,7 @@ static int emit_mocs_l3cc_table(struct i915_request *rq,
> >   /**
> >    * intel_mocs_init_l3cc_table() - program the mocs control table
> > - * @dev_priv:      i915 device private
> > + * @i915:      i915 device private
> >    *
> >    * This function simply programs the mocs registers for the given table
> >    * starting at the given address. This register set is  programmed in pairs.
> > @@ -494,13 +494,13 @@ static int emit_mocs_l3cc_table(struct i915_request *rq,
> >    *
> >    * Return: Nothing.
> >    */
> > -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv)
> > +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_mocs_table table;
> >   	unsigned int i;
> >   	u16 unused_value;
> > -	if (!get_mocs_settings(dev_priv, &table))
> > +	if (!get_mocs_settings(i915, &table))
> >   		return;
> >   	/* Set unused values to PTE */
> > diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.h b/drivers/gpu/drm/i915/gt/intel_mocs.h
> > index 0913704a1af2..6f8973fbcb3e 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_mocs.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_mocs.h
> > @@ -54,7 +54,7 @@ struct i915_request;
> >   struct intel_engine_cs;
> >   int intel_rcs_context_init_mocs(struct i915_request *rq);
> > -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv);
> > +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915);
> >   void intel_mocs_init_engine(struct intel_engine_cs *engine);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/gt/intel_reset.c b/drivers/gpu/drm/i915/gt/intel_reset.c
> > index 6368b37f26d1..d5cd3cccb407 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_reset.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_reset.c
> > @@ -1208,13 +1208,13 @@ static void clear_error_registers(struct drm_i915_private *i915,
> >   	}
> >   }
> > -static void gen6_check_faults(struct drm_i915_private *dev_priv)
> > +static void gen6_check_faults(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	u32 fault;
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		fault = GEN6_RING_FAULT_REG_READ(engine);
> >   		if (fault & RING_FAULT_VALID) {
> >   			DRM_DEBUG_DRIVER("Unexpected fault\n"
> > @@ -1230,7 +1230,7 @@ static void gen6_check_faults(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void gen8_check_faults(struct drm_i915_private *dev_priv)
> > +static void gen8_check_faults(struct drm_i915_private *i915)
> >   {
> >   	u32 fault = I915_READ(GEN8_RING_FAULT_REG);
> > diff --git a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> > index b3bf47e8162f..8ec2d332e206 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c
> > @@ -489,11 +489,11 @@ static void set_hwstam(struct intel_engine_cs *engine, u32 mask)
> >   static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	u32 addr;
> >   	addr = lower_32_bits(phys);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		addr |= (phys >> 28) & 0xf0;
> >   	I915_WRITE(HWS_PGA, addr);
> > @@ -515,14 +515,14 @@ static void ring_setup_phys_status_page(struct intel_engine_cs *engine)
> >   static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	i915_reg_t hwsp;
> >   	/*
> >   	 * The ring status page addresses are no longer next to the rest of
> >   	 * the ring registers as of gen7.
> >   	 */
> > -	if (IS_GEN(dev_priv, 7)) {
> > +	if (IS_GEN(i915, 7)) {
> >   		switch (engine->id) {
> >   		/*
> >   		 * No more rings exist on Gen7. Default case is only to shut up
> > @@ -544,7 +544,7 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
> >   			hwsp = VEBOX_HWS_PGA_GEN7;
> >   			break;
> >   		}
> > -	} else if (IS_GEN(dev_priv, 6)) {
> > +	} else if (IS_GEN(i915, 6)) {
> >   		hwsp = RING_HWS_PGA_GEN6(engine->mmio_base);
> >   	} else {
> >   		hwsp = RING_HWS_PGA(engine->mmio_base);
> > @@ -556,9 +556,9 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset)
> >   static void flush_cs_tlb(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> > -	if (!IS_GEN_RANGE(dev_priv, 6, 7))
> > +	if (!IS_GEN_RANGE(i915, 6, 7))
> >   		return;
> >   	/* ring should be idle before issuing a sync flush*/
> > @@ -585,9 +585,9 @@ static void ring_setup_status_page(struct intel_engine_cs *engine)
> >   static bool stop_ring(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> > -	if (INTEL_GEN(dev_priv) > 2) {
> > +	if (INTEL_GEN(i915) > 2) {
> >   		ENGINE_WRITE(engine,
> >   			     RING_MI_MODE, _MASKED_BIT_ENABLE(STOP_RING));
> >   		if (intel_wait_for_register(engine->uncore,
> > @@ -622,7 +622,7 @@ static bool stop_ring(struct intel_engine_cs *engine)
> >   static int xcs_resume(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	struct intel_ring *ring = engine->buffer;
> >   	int ret = 0;
> > @@ -654,7 +654,7 @@ static int xcs_resume(struct intel_engine_cs *engine)
> >   		}
> >   	}
> > -	if (HWS_NEEDS_PHYSICAL(dev_priv))
> > +	if (HWS_NEEDS_PHYSICAL(i915))
> >   		ring_setup_phys_status_page(engine);
> >   	else
> >   		ring_setup_status_page(engine);
> > @@ -705,7 +705,7 @@ static int xcs_resume(struct intel_engine_cs *engine)
> >   		goto out;
> >   	}
> > -	if (INTEL_GEN(dev_priv) > 2)
> > +	if (INTEL_GEN(i915) > 2)
> >   		ENGINE_WRITE(engine,
> >   			     RING_MI_MODE, _MASKED_BIT_DISABLE(STOP_RING));
> > @@ -815,7 +815,7 @@ static int intel_rcs_ctx_init(struct i915_request *rq)
> >   static int rcs_resume(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> >   	/*
> >   	 * Disable CONSTANT_BUFFER before it is loaded from the context
> > @@ -827,12 +827,12 @@ static int rcs_resume(struct intel_engine_cs *engine)
> >   	 * they are already accustomed to from before contexts were
> >   	 * enabled.
> >   	 */
> > -	if (IS_GEN(dev_priv, 4))
> > +	if (IS_GEN(i915, 4))
> >   		I915_WRITE(ECOSKPD,
> >   			   _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE));
> >   	/* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */
> > -	if (IS_GEN_RANGE(dev_priv, 4, 6))
> > +	if (IS_GEN_RANGE(i915, 4, 6))
> >   		I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH));
> >   	/* We need to disable the AsyncFlip performance optimisations in order
> > @@ -841,22 +841,22 @@ static int rcs_resume(struct intel_engine_cs *engine)
> >   	 *
> >   	 * WaDisableAsyncFlipPerfMode:snb,ivb,hsw,vlv
> >   	 */
> > -	if (IS_GEN_RANGE(dev_priv, 6, 7))
> > +	if (IS_GEN_RANGE(i915, 6, 7))
> >   		I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(ASYNC_FLIP_PERF_DISABLE));
> >   	/* Required for the hardware to program scanline values for waiting */
> >   	/* WaEnableFlushTlbInvalidationMode:snb */
> > -	if (IS_GEN(dev_priv, 6))
> > +	if (IS_GEN(i915, 6))
> >   		I915_WRITE(GFX_MODE,
> >   			   _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT));
> >   	/* WaBCSVCSTlbInvalidationMode:ivb,vlv,hsw */
> > -	if (IS_GEN(dev_priv, 7))
> > +	if (IS_GEN(i915, 7))
> >   		I915_WRITE(GFX_MODE_GEN7,
> >   			   _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT) |
> >   			   _MASKED_BIT_ENABLE(GFX_REPLAY_MODE));
> > -	if (IS_GEN(dev_priv, 6)) {
> > +	if (IS_GEN(i915, 6)) {
> >   		/* From the Sandybridge PRM, volume 1 part 3, page 24:
> >   		 * "If this bit is set, STCunit will have LRA as replacement
> >   		 *  policy. [...] This bit must be reset.  LRA replacement
> > @@ -866,7 +866,7 @@ static int rcs_resume(struct intel_engine_cs *engine)
> >   			   _MASKED_BIT_DISABLE(CM0_STC_EVICT_DISABLE_LRA_SNB));
> >   	}
> > -	if (IS_GEN_RANGE(dev_priv, 6, 7))
> > +	if (IS_GEN_RANGE(i915, 6, 7))
> >   		I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING));
> >   	return xcs_resume(engine);
> > @@ -1227,15 +1227,15 @@ void intel_ring_unpin(struct intel_ring *ring)
> >   }
> >   static struct i915_vma *
> > -intel_ring_create_vma(struct drm_i915_private *dev_priv, int size)
> > +intel_ring_create_vma(struct drm_i915_private *i915, int size)
> >   {
> > -	struct i915_address_space *vm = &dev_priv->ggtt.vm;
> > +	struct i915_address_space *vm = &i915->ggtt.vm;
> >   	struct drm_i915_gem_object *obj;
> >   	struct i915_vma *vma;
> > -	obj = i915_gem_object_create_stolen(dev_priv, size);
> > +	obj = i915_gem_object_create_stolen(i915, size);
> >   	if (!obj)
> > -		obj = i915_gem_object_create_internal(dev_priv, size);
> > +		obj = i915_gem_object_create_internal(i915, size);
> >   	if (IS_ERR(obj))
> >   		return ERR_CAST(obj);
> > @@ -2115,9 +2115,9 @@ static void gen6_bsd_set_default_submission(struct intel_engine_cs *engine)
> >   static void ring_destroy(struct intel_engine_cs *engine)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> > -	WARN_ON(INTEL_GEN(dev_priv) > 2 &&
> > +	WARN_ON(INTEL_GEN(i915) > 2 &&
> >   		(ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE) == 0);
> >   	intel_ring_unpin(engine->buffer);
> > diff --git a/drivers/gpu/drm/i915/gvt/aperture_gm.c b/drivers/gpu/drm/i915/gvt/aperture_gm.c
> > index 1fa2f65c3cd1..8807de566361 100644
> > --- a/drivers/gpu/drm/i915/gvt/aperture_gm.c
> > +++ b/drivers/gpu/drm/i915/gvt/aperture_gm.c
> > @@ -40,7 +40,7 @@
> >   static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	unsigned int flags;
> >   	u64 start, end, size;
> >   	struct drm_mm_node *node;
> > @@ -60,14 +60,14 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> >   		flags = PIN_MAPPABLE;
> >   	}
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	mmio_hw_access_pre(dev_priv);
> > -	ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node,
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	mmio_hw_access_pre(i915);
> > +	ret = i915_gem_gtt_insert(&i915->ggtt.vm, node,
> >   				  size, I915_GTT_PAGE_SIZE,
> >   				  I915_COLOR_UNEVICTABLE,
> >   				  start, end, flags);
> > -	mmio_hw_access_post(dev_priv);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mmio_hw_access_post(i915);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	if (ret)
> >   		gvt_err("fail to alloc %s gm space from host\n",
> >   			high_gm ? "high" : "low");
> > @@ -78,7 +78,7 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm)
> >   static int alloc_vgpu_gm(struct intel_vgpu *vgpu)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ret;
> >   	ret = alloc_gm(vgpu, false);
> > @@ -97,20 +97,20 @@ static int alloc_vgpu_gm(struct intel_vgpu *vgpu)
> >   	return 0;
> >   out_free_aperture:
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	drm_mm_remove_node(&vgpu->gm.low_gm_node);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return ret;
> >   }
> >   static void free_vgpu_gm(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	drm_mm_remove_node(&vgpu->gm.low_gm_node);
> >   	drm_mm_remove_node(&vgpu->gm.high_gm_node);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   }
> >   /**
> > @@ -127,11 +127,11 @@ void intel_vgpu_write_fence(struct intel_vgpu *vgpu,
> >   		u32 fence, u64 value)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	struct drm_i915_fence_reg *reg;
> >   	i915_reg_t fence_reg_lo, fence_reg_hi;
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	assert_rpm_wakelock_held(i915);
> >   	if (WARN_ON(fence >= vgpu_fence_sz(vgpu)))
> >   		return;
> > @@ -162,41 +162,41 @@ static void _clear_vgpu_fence(struct intel_vgpu *vgpu)
> >   static void free_vgpu_fence(struct intel_vgpu *vgpu)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	struct drm_i915_fence_reg *reg;
> >   	u32 i;
> >   	if (WARN_ON(!vgpu_fence_sz(vgpu)))
> >   		return;
> > -	intel_runtime_pm_get(dev_priv);
> > +	intel_runtime_pm_get(i915);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	_clear_vgpu_fence(vgpu);
> >   	for (i = 0; i < vgpu_fence_sz(vgpu); i++) {
> >   		reg = vgpu->fence.regs[i];
> >   		i915_unreserve_fence(reg);
> >   		vgpu->fence.regs[i] = NULL;
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   }
> >   static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	struct drm_i915_fence_reg *reg;
> >   	int i;
> > -	intel_runtime_pm_get(dev_priv);
> > +	intel_runtime_pm_get(i915);
> >   	/* Request fences from host */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	for (i = 0; i < vgpu_fence_sz(vgpu); i++) {
> > -		reg = i915_reserve_fence(dev_priv);
> > +		reg = i915_reserve_fence(i915);
> >   		if (IS_ERR(reg))
> >   			goto out_free_fence;
> > @@ -205,8 +205,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
> >   	_clear_vgpu_fence(vgpu);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   	return 0;
> >   out_free_fence:
> >   	gvt_vgpu_err("Failed to alloc fences\n");
> > @@ -218,8 +218,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu)
> >   		i915_unreserve_fence(reg);
> >   		vgpu->fence.regs[i] = NULL;
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   	return -ENOSPC;
> >   }
> > @@ -313,11 +313,11 @@ void intel_vgpu_free_resource(struct intel_vgpu *vgpu)
> >    */
> >   void intel_vgpu_reset_resource(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > -	intel_runtime_pm_get(dev_priv);
> > +	intel_runtime_pm_get(i915);
> >   	_clear_vgpu_fence(vgpu);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   }
> >   /**
> > diff --git a/drivers/gpu/drm/i915/gvt/cfg_space.c b/drivers/gpu/drm/i915/gvt/cfg_space.c
> > index 19cf1bbe059d..d3a744e7c959 100644
> > --- a/drivers/gpu/drm/i915/gvt/cfg_space.c
> > +++ b/drivers/gpu/drm/i915/gvt/cfg_space.c
> > @@ -391,9 +391,9 @@ void intel_vgpu_init_cfg_space(struct intel_vgpu *vgpu,
> >   	memset(vgpu_cfg_space(vgpu) + INTEL_GVT_PCI_OPREGION, 0, 4);
> >   	vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_GTTMMIO].size =
> > -				pci_resource_len(gvt->dev_priv->drm.pdev, 0);
> > +				pci_resource_len(gvt->i915->drm.pdev, 0);
> >   	vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_APERTURE].size =
> > -				pci_resource_len(gvt->dev_priv->drm.pdev, 2);
> > +				pci_resource_len(gvt->i915->drm.pdev, 2);
> >   	memset(vgpu_cfg_space(vgpu) + PCI_ROM_ADDRESS, 0, 4);
> >   }
> > diff --git a/drivers/gpu/drm/i915/gvt/cmd_parser.c b/drivers/gpu/drm/i915/gvt/cmd_parser.c
> > index 6ea88270c818..2b09acc0502f 100644
> > --- a/drivers/gpu/drm/i915/gvt/cmd_parser.c
> > +++ b/drivers/gpu/drm/i915/gvt/cmd_parser.c
> > @@ -819,7 +819,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s,
> >   	unsigned int data;
> >   	u32 ring_base;
> >   	u32 nopid;
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> >   	if (!strcmp(cmd, "lri"))
> >   		data = cmd_val(s, index + 1);
> > @@ -829,7 +829,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s,
> >   		return -EINVAL;
> >   	}
> > -	ring_base = dev_priv->engine[s->ring_id]->mmio_base;
> > +	ring_base = i915->engine[s->ring_id]->mmio_base;
> >   	nopid = i915_mmio_reg_offset(RING_NOPID(ring_base));
> >   	if (!intel_gvt_in_force_nonpriv_whitelist(gvt, data) &&
> > @@ -905,7 +905,7 @@ static int cmd_reg_handler(struct parser_exec_state *s,
> >   	 * update reg values in it into vregs, so LRIs in workload with
> >   	 * inhibit context will restore with correct values
> >   	 */
> > -	if (IS_GEN(gvt->dev_priv, 9) &&
> > +	if (IS_GEN(gvt->i915, 9) &&
> >   			intel_gvt_mmio_is_in_ctx(gvt, offset) &&
> >   			!strncmp(cmd, "lri", 3)) {
> >   		intel_gvt_hypervisor_read_gpa(s->vgpu,
> > @@ -946,7 +946,7 @@ static int cmd_handler_lri(struct parser_exec_state *s)
> >   	struct intel_gvt *gvt = s->vgpu->gvt;
> >   	for (i = 1; i < cmd_len; i += 2) {
> > -		if (IS_BROADWELL(gvt->dev_priv) && s->ring_id != RCS0) {
> > +		if (IS_BROADWELL(gvt->i915) && s->ring_id != RCS0) {
> >   			if (s->ring_id == BCS0 &&
> >   			    cmd_reg(s, i) == i915_mmio_reg_offset(DERRMR))
> >   				ret |= 0;
> > @@ -968,7 +968,7 @@ static int cmd_handler_lrr(struct parser_exec_state *s)
> >   	int cmd_len = cmd_length(s);
> >   	for (i = 1; i < cmd_len; i += 2) {
> > -		if (IS_BROADWELL(s->vgpu->gvt->dev_priv))
> > +		if (IS_BROADWELL(s->vgpu->gvt->i915))
> >   			ret |= ((cmd_reg_inhibit(s, i) ||
> >   					(cmd_reg_inhibit(s, i + 1)))) ?
> >   				-EBADRQC : 0;
> > @@ -996,7 +996,7 @@ static int cmd_handler_lrm(struct parser_exec_state *s)
> >   	int cmd_len = cmd_length(s);
> >   	for (i = 1; i < cmd_len;) {
> > -		if (IS_BROADWELL(gvt->dev_priv))
> > +		if (IS_BROADWELL(gvt->i915))
> >   			ret |= (cmd_reg_inhibit(s, i)) ? -EBADRQC : 0;
> >   		if (ret)
> >   			break;
> > @@ -1180,7 +1180,7 @@ struct plane_code_mapping {
> >   static int gen8_decode_mi_display_flip(struct parser_exec_state *s,
> >   		struct mi_display_flip_command_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> >   	struct plane_code_mapping gen8_plane_code[] = {
> >   		[0] = {PIPE_A, PLANE_A, PRIMARY_A_FLIP_DONE},
> >   		[1] = {PIPE_B, PLANE_A, PRIMARY_B_FLIP_DONE},
> > @@ -1226,7 +1226,7 @@ static int gen8_decode_mi_display_flip(struct parser_exec_state *s,
> >   static int skl_decode_mi_display_flip(struct parser_exec_state *s,
> >   		struct mi_display_flip_command_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> >   	struct intel_vgpu *vgpu = s->vgpu;
> >   	u32 dword0 = cmd_val(s, 0);
> >   	u32 dword1 = cmd_val(s, 1);
> > @@ -1285,13 +1285,13 @@ static int skl_decode_mi_display_flip(struct parser_exec_state *s,
> >   static int gen8_check_mi_display_flip(struct parser_exec_state *s,
> >   		struct mi_display_flip_command_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> >   	u32 stride, tile;
> >   	if (!info->async_flip)
> >   		return 0;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		stride = vgpu_vreg_t(s->vgpu, info->stride_reg) & GENMASK(9, 0);
> >   		tile = (vgpu_vreg_t(s->vgpu, info->ctrl_reg) &
> >   				GENMASK(12, 10)) >> 10;
> > @@ -1314,12 +1314,12 @@ static int gen8_update_plane_mmio_from_mi_display_flip(
> >   		struct parser_exec_state *s,
> >   		struct mi_display_flip_command_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> >   	struct intel_vgpu *vgpu = s->vgpu;
> >   	set_mask_bits(&vgpu_vreg_t(vgpu, info->surf_reg), GENMASK(31, 12),
> >   		      info->surf_val << 12);
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		set_mask_bits(&vgpu_vreg_t(vgpu, info->stride_reg), GENMASK(9, 0),
> >   			      info->stride_val);
> >   		set_mask_bits(&vgpu_vreg_t(vgpu, info->ctrl_reg), GENMASK(12, 10),
> > @@ -1345,11 +1345,11 @@ static int gen8_update_plane_mmio_from_mi_display_flip(
> >   static int decode_mi_display_flip(struct parser_exec_state *s,
> >   		struct mi_display_flip_command_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = s->vgpu->gvt->i915;
> > -	if (IS_BROADWELL(dev_priv))
> > +	if (IS_BROADWELL(i915))
> >   		return gen8_decode_mi_display_flip(s, info);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return skl_decode_mi_display_flip(s, info);
> >   	return -ENODEV;
> > @@ -1756,7 +1756,7 @@ static int perform_bb_shadow(struct parser_exec_state *s)
> >   	if (bb->ppgtt)
> >   		start_offset = gma & ~I915_GTT_PAGE_MASK;
> > -	bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->dev_priv,
> > +	bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->i915,
> >   					       round_up(bb_size + start_offset,
> >   							PAGE_SIZE));
> >   	if (IS_ERR(bb->obj)) {
> > @@ -2830,7 +2830,7 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx)
> >   	int ret = 0;
> >   	void *map;
> > -	obj = i915_gem_object_create_shmem(workload->vgpu->gvt->dev_priv,
> > +	obj = i915_gem_object_create_shmem(workload->vgpu->gvt->i915,
> >   					   roundup(ctx_size + CACHELINE_BYTES,
> >   						   PAGE_SIZE));
> >   	if (IS_ERR(obj))
> > diff --git a/drivers/gpu/drm/i915/gvt/debugfs.c b/drivers/gpu/drm/i915/gvt/debugfs.c
> > index 2fb7b73b260d..1ff7f11c89a2 100644
> > --- a/drivers/gpu/drm/i915/gvt/debugfs.c
> > +++ b/drivers/gpu/drm/i915/gvt/debugfs.c
> > @@ -58,7 +58,7 @@ static int mmio_offset_compare(void *priv,
> >   static inline int mmio_diff_handler(struct intel_gvt *gvt,
> >   				    u32 offset, void *data)
> >   {
> > -	struct drm_i915_private *i915 = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	struct mmio_diff_param *param = data;
> >   	struct diff_mmio *node;
> >   	u32 preg, vreg;
> > @@ -98,10 +98,10 @@ static int vgpu_mmio_diff_show(struct seq_file *s, void *unused)
> >   	mutex_lock(&gvt->lock);
> >   	spin_lock_bh(&gvt->scheduler.mmio_context_lock);
> > -	mmio_hw_access_pre(gvt->dev_priv);
> > +	mmio_hw_access_pre(gvt->i915);
> >   	/* Recognize all the diff mmios to list. */
> >   	intel_gvt_for_each_tracked_mmio(gvt, mmio_diff_handler, &param);
> > -	mmio_hw_access_post(gvt->dev_priv);
> > +	mmio_hw_access_post(gvt->i915);
> >   	spin_unlock_bh(&gvt->scheduler.mmio_context_lock);
> >   	mutex_unlock(&gvt->lock);
> > @@ -142,7 +142,7 @@ static int
> >   vgpu_scan_nonprivbb_set(void *data, u64 val)
> >   {
> >   	struct intel_vgpu *vgpu = (struct intel_vgpu *)data;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	enum intel_engine_id id;
> >   	char buf[128], *s;
> >   	int len;
> > @@ -164,7 +164,7 @@ vgpu_scan_nonprivbb_set(void *data, u64 val)
> >   	for (id = 0; id < I915_NUM_ENGINES; id++) {
> >   		struct intel_engine_cs *engine;
> > -		engine = dev_priv->engine[id];
> > +		engine = i915->engine[id];
> >   		if (engine && (val & (1 << id))) {
> >   			len = snprintf(s, 4, "%d, ", engine->id);
> >   			s += len;
> > @@ -240,7 +240,7 @@ void intel_gvt_debugfs_remove_vgpu(struct intel_vgpu *vgpu)
> >    */
> >   int intel_gvt_debugfs_init(struct intel_gvt *gvt)
> >   {
> > -	struct drm_minor *minor = gvt->dev_priv->drm.primary;
> > +	struct drm_minor *minor = gvt->i915->drm.primary;
> >   	struct dentry *ent;
> >   	gvt->debugfs_root = debugfs_create_dir("gvt", minor->debugfs_root);
> > diff --git a/drivers/gpu/drm/i915/gvt/display.c b/drivers/gpu/drm/i915/gvt/display.c
> > index e1c313da6c00..b037aeed2f40 100644
> > --- a/drivers/gpu/drm/i915/gvt/display.c
> > +++ b/drivers/gpu/drm/i915/gvt/display.c
> > @@ -57,7 +57,7 @@ static int get_edp_pipe(struct intel_vgpu *vgpu)
> >   static int edp_pipe_is_enabled(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	if (!(vgpu_vreg_t(vgpu, PIPECONF(_PIPE_EDP)) & PIPECONF_ENABLE))
> >   		return 0;
> > @@ -69,7 +69,7 @@ static int edp_pipe_is_enabled(struct intel_vgpu *vgpu)
> >   int pipe_is_enabled(struct intel_vgpu *vgpu, int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	if (WARN_ON(pipe < PIPE_A || pipe >= I915_MAX_PIPES))
> >   		return -EINVAL;
> > @@ -168,10 +168,10 @@ static u8 dpcd_fix_data[DPCD_HEADER_SIZE] = {
> >   static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int pipe;
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) &= ~(BXT_DE_PORT_HP_DDIA |
> >   			BXT_DE_PORT_HP_DDIB |
> >   			BXT_DE_PORT_HP_DDIC);
> > @@ -198,8 +198,8 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   			SDE_PORTC_HOTPLUG_CPT |
> >   			SDE_PORTD_HOTPLUG_CPT);
> > -	if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> > -	    IS_COFFEELAKE(dev_priv)) {
> > +	if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> > +	    IS_COFFEELAKE(i915)) {
> >   		vgpu_vreg_t(vgpu, SDEISR) &= ~(SDE_PORTA_HOTPLUG_SPT |
> >   				SDE_PORTE_HOTPLUG_SPT);
> >   		vgpu_vreg_t(vgpu, SKL_FUSE_STATUS) |=
> > @@ -223,7 +223,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   			(TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> >   			(PORT_B << TRANS_DDI_PORT_SHIFT) |
> >   			TRANS_DDI_FUNC_ENABLE);
> > -		if (IS_BROADWELL(dev_priv)) {
> > +		if (IS_BROADWELL(i915)) {
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) &=
> >   				~PORT_CLK_SEL_MASK;
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) |=
> > @@ -243,7 +243,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   			(TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> >   			(PORT_C << TRANS_DDI_PORT_SHIFT) |
> >   			TRANS_DDI_FUNC_ENABLE);
> > -		if (IS_BROADWELL(dev_priv)) {
> > +		if (IS_BROADWELL(i915)) {
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) &=
> >   				~PORT_CLK_SEL_MASK;
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) |=
> > @@ -263,7 +263,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   			(TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI |
> >   			(PORT_D << TRANS_DDI_PORT_SHIFT) |
> >   			TRANS_DDI_FUNC_ENABLE);
> > -		if (IS_BROADWELL(dev_priv)) {
> > +		if (IS_BROADWELL(i915)) {
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) &=
> >   				~PORT_CLK_SEL_MASK;
> >   			vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) |=
> > @@ -274,14 +274,14 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   		vgpu_vreg_t(vgpu, SFUSE_STRAP) |= SFUSE_STRAP_DDID_DETECTED;
> >   	}
> > -	if ((IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> > -	     IS_COFFEELAKE(dev_priv)) &&
> > +	if ((IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> > +	     IS_COFFEELAKE(i915)) &&
> >   			intel_vgpu_has_monitor_on_port(vgpu, PORT_E)) {
> >   		vgpu_vreg_t(vgpu, SDEISR) |= SDE_PORTE_HOTPLUG_SPT;
> >   	}
> >   	if (intel_vgpu_has_monitor_on_port(vgpu, PORT_A)) {
> > -		if (IS_BROADWELL(dev_priv))
> > +		if (IS_BROADWELL(i915))
> >   			vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) |=
> >   				GEN8_PORT_DP_A_HOTPLUG;
> >   		else
> > @@ -291,11 +291,11 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu)
> >   	}
> >   	/* Clear host CRT status, so guest couldn't detect this host CRT. */
> > -	if (IS_BROADWELL(dev_priv))
> > +	if (IS_BROADWELL(i915))
> >   		vgpu_vreg_t(vgpu, PCH_ADPA) &= ~ADPA_CRT_HOTPLUG_MONITOR_MASK;
> >   	/* Disable Primary/Sprite/Cursor plane */
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		vgpu_vreg_t(vgpu, DSPCNTR(pipe)) &= ~DISPLAY_PLANE_ENABLE;
> >   		vgpu_vreg_t(vgpu, SPRCTL(pipe)) &= ~SPRITE_ENABLE;
> >   		vgpu_vreg_t(vgpu, CURCNTR(pipe)) &= ~MCURSOR_MODE;
> > @@ -389,7 +389,7 @@ void intel_gvt_check_vblank_emulation(struct intel_gvt *gvt)
> >   static void emulate_vblank_on_pipe(struct intel_vgpu *vgpu, int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_vgpu_irq *irq = &vgpu->irq;
> >   	int vblank_event[] = {
> >   		[PIPE_A] = PIPE_A_VBLANK,
> > @@ -421,7 +421,7 @@ static void emulate_vblank(struct intel_vgpu *vgpu)
> >   	int pipe;
> >   	mutex_lock(&vgpu->vgpu_lock);
> > -	for_each_pipe(vgpu->gvt->dev_priv, pipe)
> > +	for_each_pipe(vgpu->gvt->i915, pipe)
> >   		emulate_vblank_on_pipe(vgpu, pipe);
> >   	mutex_unlock(&vgpu->vgpu_lock);
> >   }
> > @@ -454,10 +454,10 @@ void intel_gvt_emulate_vblank(struct intel_gvt *gvt)
> >    */
> >   void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	/* TODO: add more platforms support */
> > -	if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv)) {
> > +	if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915)) {
> >   		if (connected) {
> >   			vgpu_vreg_t(vgpu, SFUSE_STRAP) |=
> >   				SFUSE_STRAP_DDID_DETECTED;
> > @@ -483,10 +483,10 @@ void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected)
> >    */
> >   void intel_vgpu_clean_display(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > -	if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> > -	    IS_COFFEELAKE(dev_priv))
> > +	if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> > +	    IS_COFFEELAKE(i915))
> >   		clean_virtual_dp_monitor(vgpu, PORT_D);
> >   	else
> >   		clean_virtual_dp_monitor(vgpu, PORT_B);
> > @@ -505,12 +505,12 @@ void intel_vgpu_clean_display(struct intel_vgpu *vgpu)
> >    */
> >   int intel_vgpu_init_display(struct intel_vgpu *vgpu, u64 resolution)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	intel_vgpu_init_i2c_edid(vgpu);
> > -	if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
> > -	    IS_COFFEELAKE(dev_priv))
> > +	if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) ||
> > +	    IS_COFFEELAKE(i915))
> >   		return setup_virtual_dp_monitor(vgpu, PORT_D, GVT_DP_D,
> >   						resolution);
> >   	else
> > diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c
> > index 41c8ebc60c63..523038d224a1 100644
> > --- a/drivers/gpu/drm/i915/gvt/dmabuf.c
> > +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c
> > @@ -39,7 +39,7 @@
> >   static int vgpu_gem_get_pages(
> >   		struct drm_i915_gem_object *obj)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct sg_table *st;
> >   	struct scatterlist *sg;
> >   	int i, ret;
> > @@ -61,7 +61,7 @@ static int vgpu_gem_get_pages(
> >   		kfree(st);
> >   		return ret;
> >   	}
> > -	gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm +
> > +	gtt_entries = (gen8_pte_t __iomem *)i915->ggtt.gsm +
> >   		(fb_info->start >> PAGE_SHIFT);
> >   	for_each_sg(st->sgl, sg, page_num, i) {
> >   		sg->offset = 0;
> > @@ -152,7 +152,7 @@ static const struct drm_i915_gem_object_ops intel_vgpu_gem_ops = {
> >   static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev,
> >   		struct intel_vgpu_fb_info *info)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_object *obj;
> >   	obj = i915_gem_object_alloc();
> > @@ -165,7 +165,7 @@ static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev,
> >   	obj->read_domains = I915_GEM_DOMAIN_GTT;
> >   	obj->write_domain = 0;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		unsigned int tiling_mode = 0;
> >   		unsigned int stride = 0;
> > @@ -360,7 +360,7 @@ static void update_fb_info(struct vfio_device_gfx_plane_info *gvt_dmabuf,
> >   int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args)
> >   {
> > -	struct drm_device *dev = &vgpu->gvt->dev_priv->drm;
> > +	struct drm_device *dev = &vgpu->gvt->i915->drm;
> >   	struct vfio_device_gfx_plane_info *gfx_plane_info = args;
> >   	struct intel_vgpu_dmabuf_obj *dmabuf_obj;
> >   	struct intel_vgpu_fb_info fb_info;
> > @@ -466,7 +466,7 @@ int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args)
> >   /* To associate an exposed dmabuf with the dmabuf_obj */
> >   int intel_vgpu_get_dmabuf(struct intel_vgpu *vgpu, unsigned int dmabuf_id)
> >   {
> > -	struct drm_device *dev = &vgpu->gvt->dev_priv->drm;
> > +	struct drm_device *dev = &vgpu->gvt->i915->drm;
> >   	struct intel_vgpu_dmabuf_obj *dmabuf_obj;
> >   	struct drm_i915_gem_object *obj;
> >   	struct dma_buf *dmabuf;
> > diff --git a/drivers/gpu/drm/i915/gvt/edid.c b/drivers/gpu/drm/i915/gvt/edid.c
> > index 1fe6124918f1..9e45cbac5eed 100644
> > --- a/drivers/gpu/drm/i915/gvt/edid.c
> > +++ b/drivers/gpu/drm/i915/gvt/edid.c
> > @@ -135,7 +135,7 @@ static void reset_gmbus_controller(struct intel_vgpu *vgpu)
> >   static int gmbus0_mmio_write(struct intel_vgpu *vgpu,
> >   			unsigned int offset, void *p_data, unsigned int bytes)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int port, pin_select;
> >   	memcpy(&vgpu_vreg(vgpu, offset), p_data, bytes);
> > @@ -147,9 +147,9 @@ static int gmbus0_mmio_write(struct intel_vgpu *vgpu,
> >   	if (pin_select == 0)
> >   		return 0;
> > -	if (IS_BROXTON(dev_priv))
> > +	if (IS_BROXTON(i915))
> >   		port = bxt_get_port_from_gmbus0(pin_select);
> > -	else if (IS_COFFEELAKE(dev_priv))
> > +	else if (IS_COFFEELAKE(i915))
> >   		port = cnp_get_port_from_gmbus0(pin_select);
> >   	else
> >   		port = get_port_from_gmbus0(pin_select);
> > diff --git a/drivers/gpu/drm/i915/gvt/execlist.c b/drivers/gpu/drm/i915/gvt/execlist.c
> > index f21b8fb5b37e..7a7a925565a0 100644
> > --- a/drivers/gpu/drm/i915/gvt/execlist.c
> > +++ b/drivers/gpu/drm/i915/gvt/execlist.c
> > @@ -40,7 +40,7 @@
> >   #define _EL_OFFSET_STATUS_PTR   0x3A0
> >   #define execlist_ring_mmio(gvt, ring_id, offset) \
> > -	(gvt->dev_priv->engine[ring_id]->mmio_base + (offset))
> > +	(gvt->i915->engine[ring_id]->mmio_base + (offset))
> >   #define valid_context(ctx) ((ctx)->valid)
> >   #define same_context(a, b) (((a)->context_id == (b)->context_id) && \
> > @@ -133,7 +133,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist,
> >   	u32 write_pointer;
> >   	u32 ctx_status_ptr_reg, ctx_status_buf_reg, offset;
> >   	unsigned long hwsp_gpa;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	ctx_status_ptr_reg = execlist_ring_mmio(vgpu->gvt, ring_id,
> >   			_EL_OFFSET_STATUS_PTR);
> > @@ -169,7 +169,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist,
> >   			status, 8);
> >   		intel_gvt_hypervisor_write_gpa(vgpu,
> >   			hwsp_gpa +
> > -			intel_hws_csb_write_index(dev_priv) * 4,
> > +			intel_hws_csb_write_index(i915) * 4,
> >   			&write_pointer, 4);
> >   	}
> > @@ -529,12 +529,12 @@ static void init_vgpu_execlist(struct intel_vgpu *vgpu, int ring_id)
> >   static void clean_execlist(struct intel_vgpu *vgpu,
> >   			   intel_engine_mask_t engine_mask)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_engine_cs *engine;
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> >   	intel_engine_mask_t tmp;
> > -	for_each_engine_masked(engine, dev_priv, engine_mask, tmp) {
> > +	for_each_engine_masked(engine, i915, engine_mask, tmp) {
> >   		kfree(s->ring_scan_buffer[engine->id]);
> >   		s->ring_scan_buffer[engine->id] = NULL;
> >   		s->ring_scan_buffer_size[engine->id] = 0;
> > @@ -544,11 +544,11 @@ static void clean_execlist(struct intel_vgpu *vgpu,
> >   static void reset_execlist(struct intel_vgpu *vgpu,
> >   			   intel_engine_mask_t engine_mask)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_engine_cs *engine;
> >   	intel_engine_mask_t tmp;
> > -	for_each_engine_masked(engine, dev_priv, engine_mask, tmp)
> > +	for_each_engine_masked(engine, i915, engine_mask, tmp)
> >   		init_vgpu_execlist(vgpu, engine->id);
> >   }
> > diff --git a/drivers/gpu/drm/i915/gvt/fb_decoder.c b/drivers/gpu/drm/i915/gvt/fb_decoder.c
> > index 65e847392aea..76c346eecda3 100644
> > --- a/drivers/gpu/drm/i915/gvt/fb_decoder.c
> > +++ b/drivers/gpu/drm/i915/gvt/fb_decoder.c
> > @@ -146,12 +146,12 @@ static int skl_format_to_drm(int format, bool rgb_order, bool alpha,
> >   static u32 intel_vgpu_get_stride(struct intel_vgpu *vgpu, int pipe,
> >   	u32 tiled, int stride_mask, int bpp)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	u32 stride_reg = vgpu_vreg_t(vgpu, DSPSTRIDE(pipe)) & stride_mask;
> >   	u32 stride = stride_reg;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		switch (tiled) {
> >   		case PLANE_CTL_TILED_LINEAR:
> >   			stride = stride_reg * 64;
> > @@ -203,7 +203,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> >   	struct intel_vgpu_primary_plane_format *plane)
> >   {
> >   	u32 val, fmt;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int pipe;
> >   	pipe = get_active_pipe(vgpu);
> > @@ -215,7 +215,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> >   	if (!plane->enabled)
> >   		return -ENODEV;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		plane->tiled = val & PLANE_CTL_TILED_MASK;
> >   		fmt = skl_format_to_drm(
> >   			val & PLANE_CTL_FORMAT_MASK,
> > @@ -256,7 +256,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu,
> >   	}
> >   	plane->stride = intel_vgpu_get_stride(vgpu, pipe, plane->tiled,
> > -		(INTEL_GEN(dev_priv) >= 9) ?
> > +		(INTEL_GEN(i915) >= 9) ?
> >   			(_PRI_PLANE_STRIDE_MASK >> 6) :
> >   				_PRI_PLANE_STRIDE_MASK, plane->bpp);
> > @@ -334,7 +334,7 @@ int intel_vgpu_decode_cursor_plane(struct intel_vgpu *vgpu,
> >   {
> >   	u32 val, mode, index;
> >   	u32 alpha_plane, alpha_force;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int pipe;
> >   	pipe = get_active_pipe(vgpu);
> > diff --git a/drivers/gpu/drm/i915/gvt/firmware.c b/drivers/gpu/drm/i915/gvt/firmware.c
> > index 049775e8e350..efdc9df4f731 100644
> > --- a/drivers/gpu/drm/i915/gvt/firmware.c
> > +++ b/drivers/gpu/drm/i915/gvt/firmware.c
> > @@ -68,7 +68,7 @@ static struct bin_attribute firmware_attr = {
> >   static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data)
> >   {
> > -	struct drm_i915_private *i915 = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	*(u32 *)(data + offset) = intel_uncore_read_notrace(&i915->uncore,
> >   							    _MMIO(offset));
> > @@ -78,7 +78,7 @@ static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data)
> >   static int expose_firmware_sysfs(struct intel_gvt *gvt)
> >   {
> >   	struct intel_gvt_device_info *info = &gvt->device_info;
> > -	struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = gvt->i915->drm.pdev;
> >   	struct gvt_firmware_header *h;
> >   	void *firmware;
> >   	void *p;
> > @@ -129,7 +129,7 @@ static int expose_firmware_sysfs(struct intel_gvt *gvt)
> >   static void clean_firmware_sysfs(struct intel_gvt *gvt)
> >   {
> > -	struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = gvt->i915->drm.pdev;
> >   	device_remove_bin_file(&pdev->dev, &firmware_attr);
> >   	vfree(firmware_attr.private);
> > @@ -153,8 +153,8 @@ static int verify_firmware(struct intel_gvt *gvt,
> >   			   const struct firmware *fw)
> >   {
> >   	struct intel_gvt_device_info *info = &gvt->device_info;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = gvt->i915;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	struct gvt_firmware_header *h;
> >   	unsigned long id, crc32_start;
> >   	const void *mem;
> > @@ -208,8 +208,8 @@ static int verify_firmware(struct intel_gvt *gvt,
> >   int intel_gvt_load_firmware(struct intel_gvt *gvt)
> >   {
> >   	struct intel_gvt_device_info *info = &gvt->device_info;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = gvt->i915;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	struct intel_gvt_firmware *firmware = &gvt->firmware;
> >   	struct gvt_firmware_header *h;
> >   	const struct firmware *fw;
> > @@ -244,7 +244,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt)
> >   	gvt_dbg_core("request hw state firmware %s...\n", path);
> > -	ret = request_firmware(&fw, path, &dev_priv->drm.pdev->dev);
> > +	ret = request_firmware(&fw, path, &i915->drm.pdev->dev);
> >   	kfree(path);
> >   	if (ret)
> > diff --git a/drivers/gpu/drm/i915/gvt/gtt.c b/drivers/gpu/drm/i915/gvt/gtt.c
> > index 53115bdae12b..390562fe3223 100644
> > --- a/drivers/gpu/drm/i915/gvt/gtt.c
> > +++ b/drivers/gpu/drm/i915/gvt/gtt.c
> > @@ -275,24 +275,24 @@ static inline int get_pse_type(int type)
> >   	return gtt_type_table[type].pse_entry_type;
> >   }
> > -static u64 read_pte64(struct drm_i915_private *dev_priv, unsigned long index)
> > +static u64 read_pte64(struct drm_i915_private *i915, unsigned long index)
> >   {
> > -	void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
> > +	void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index;
> >   	return readq(addr);
> >   }
> > -static void ggtt_invalidate(struct drm_i915_private *dev_priv)
> > +static void ggtt_invalidate(struct drm_i915_private *i915)
> >   {
> > -	mmio_hw_access_pre(dev_priv);
> > +	mmio_hw_access_pre(i915);
> >   	I915_WRITE(GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN);
> > -	mmio_hw_access_post(dev_priv);
> > +	mmio_hw_access_post(i915);
> >   }
> > -static void write_pte64(struct drm_i915_private *dev_priv,
> > +static void write_pte64(struct drm_i915_private *i915,
> >   		unsigned long index, u64 pte)
> >   {
> > -	void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
> > +	void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index;
> >   	writeq(pte, addr);
> >   }
> > @@ -315,7 +315,7 @@ static inline int gtt_get_entry64(void *pt,
> >   		if (WARN_ON(ret))
> >   			return ret;
> >   	} else if (!pt) {
> > -		e->val64 = read_pte64(vgpu->gvt->dev_priv, index);
> > +		e->val64 = read_pte64(vgpu->gvt->i915, index);
> >   	} else {
> >   		e->val64 = *((u64 *)pt + index);
> >   	}
> > @@ -340,7 +340,7 @@ static inline int gtt_set_entry64(void *pt,
> >   		if (WARN_ON(ret))
> >   			return ret;
> >   	} else if (!pt) {
> > -		write_pte64(vgpu->gvt->dev_priv, index, e->val64);
> > +		write_pte64(vgpu->gvt->i915, index, e->val64);
> >   	} else {
> >   		*((u64 *)pt + index) = e->val64;
> >   	}
> > @@ -734,7 +734,7 @@ static int detach_oos_page(struct intel_vgpu *vgpu,
> >   static void ppgtt_free_spt(struct intel_vgpu_ppgtt_spt *spt)
> >   {
> > -	struct device *kdev = &spt->vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *kdev = &spt->vgpu->gvt->i915->drm.pdev->dev;
> >   	trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type);
> > @@ -819,7 +819,7 @@ static int reclaim_one_ppgtt_mm(struct intel_gvt *gvt);
> >   static struct intel_vgpu_ppgtt_spt *ppgtt_alloc_spt(
> >   		struct intel_vgpu *vgpu, enum intel_gvt_gtt_type type)
> >   {
> > -	struct device *kdev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *kdev = &vgpu->gvt->i915->drm.pdev->dev;
> >   	struct intel_vgpu_ppgtt_spt *spt = NULL;
> >   	dma_addr_t daddr;
> >   	int ret;
> > @@ -1044,14 +1044,14 @@ static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt)
> >   static bool vgpu_ips_enabled(struct intel_vgpu *vgpu)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > -	if (INTEL_GEN(dev_priv) == 9 || INTEL_GEN(dev_priv) == 10) {
> > +	if (INTEL_GEN(i915) == 9 || INTEL_GEN(i915) == 10) {
> >   		u32 ips = vgpu_vreg_t(vgpu, GEN8_GAMW_ECO_DEV_RW_IA) &
> >   			GAMW_ECO_ENABLE_64K_IPS_FIELD;
> >   		return ips == GAMW_ECO_ENABLE_64K_IPS_FIELD;
> > -	} else if (INTEL_GEN(dev_priv) >= 11) {
> > +	} else if (INTEL_GEN(i915) >= 11) {
> >   		/* 64K paging only controlled by IPS bit in PTE now. */
> >   		return true;
> >   	} else
> > @@ -1153,7 +1153,7 @@ static int is_2MB_gtt_possible(struct intel_vgpu *vgpu,
> >   	struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops;
> >   	unsigned long pfn;
> > -	if (!HAS_PAGE_SIZES(vgpu->gvt->dev_priv, I915_GTT_PAGE_SIZE_2M))
> > +	if (!HAS_PAGE_SIZES(vgpu->gvt->i915, I915_GTT_PAGE_SIZE_2M))
> >   		return 0;
> >   	pfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, ops->get_pfn(entry));
> > @@ -2301,7 +2301,7 @@ static int emulate_ggtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> >   	ggtt_invalidate_pte(vgpu, &e);
> >   	ggtt_set_host_entry(ggtt_mm, &m, g_gtt_index);
> > -	ggtt_invalidate(gvt->dev_priv);
> > +	ggtt_invalidate(gvt->i915);
> >   	return 0;
> >   }
> > @@ -2340,7 +2340,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu,
> >   				vgpu->gvt->device_info.gtt_entry_size_shift;
> >   	void *scratch_pt;
> >   	int i;
> > -	struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> >   	dma_addr_t daddr;
> >   	if (WARN_ON(type < GTT_TYPE_PPGTT_PTE_PT || type >= GTT_TYPE_MAX))
> > @@ -2397,7 +2397,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu,
> >   static int release_scratch_page_tree(struct intel_vgpu *vgpu)
> >   {
> >   	int i;
> > -	struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> >   	dma_addr_t daddr;
> >   	for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) {
> > @@ -2669,7 +2669,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt)
> >   {
> >   	int ret;
> >   	void *page;
> > -	struct device *dev = &gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &gvt->i915->drm.pdev->dev;
> >   	dma_addr_t daddr;
> >   	gvt_dbg_core("init gtt\n");
> > @@ -2718,7 +2718,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt)
> >    */
> >   void intel_gvt_clean_gtt(struct intel_gvt *gvt)
> >   {
> > -	struct device *dev = &gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &gvt->i915->drm.pdev->dev;
> >   	dma_addr_t daddr = (dma_addr_t)(gvt->gtt.scratch_mfn <<
> >   					I915_GTT_PAGE_SHIFT);
> > @@ -2766,7 +2766,7 @@ void intel_vgpu_invalidate_ppgtt(struct intel_vgpu *vgpu)
> >   void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	struct intel_gvt_gtt_pte_ops *pte_ops = vgpu->gvt->gtt.pte_ops;
> >   	struct intel_gvt_gtt_entry entry = {.type = GTT_TYPE_GGTT_PTE};
> >   	struct intel_gvt_gtt_entry old_entry;
> > @@ -2796,7 +2796,7 @@ void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old)
> >   		ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++);
> >   	}
> > -	ggtt_invalidate(dev_priv);
> > +	ggtt_invalidate(i915);
> >   }
> >   /**
> > diff --git a/drivers/gpu/drm/i915/gvt/gvt.c b/drivers/gpu/drm/i915/gvt/gvt.c
> > index 43f4242062dd..b84cdd8d8c98 100644
> > --- a/drivers/gpu/drm/i915/gvt/gvt.c
> > +++ b/drivers/gpu/drm/i915/gvt/gvt.c
> > @@ -52,7 +52,7 @@ static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt,
> >   	int i;
> >   	struct intel_vgpu_type *t;
> >   	const char *driver_name = dev_driver_string(
> > -			&gvt->dev_priv->drm.pdev->dev);
> > +			&gvt->i915->drm.pdev->dev);
> >   	for (i = 0; i < gvt->num_types; i++) {
> >   		t = &gvt->types[i];
> > @@ -191,7 +191,7 @@ static const struct intel_gvt_ops intel_gvt_ops = {
> >   static void init_device_info(struct intel_gvt *gvt)
> >   {
> >   	struct intel_gvt_device_info *info = &gvt->device_info;
> > -	struct pci_dev *pdev = gvt->dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = gvt->i915->drm.pdev;
> >   	info->max_support_vgpus = 8;
> >   	info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE;
> > @@ -257,15 +257,15 @@ static int init_service_thread(struct intel_gvt *gvt)
> >   /**
> >    * intel_gvt_clean_device - clean a GVT device
> > - * @dev_priv: i915 private
> > + * @i915: i915 private
> >    *
> >    * This function is called at the driver unloading stage, to free the
> >    * resources owned by a GVT device.
> >    *
> >    */
> > -void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> > +void intel_gvt_clean_device(struct drm_i915_private *i915)
> >   {
> > -	struct intel_gvt *gvt = to_gvt(dev_priv);
> > +	struct intel_gvt *gvt = to_gvt(i915);
> >   	if (WARN_ON(!gvt))
> >   		return;
> > @@ -285,13 +285,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> >   	intel_gvt_clean_mmio_info(gvt);
> >   	idr_destroy(&gvt->vgpu_idr);
> > -	kfree(dev_priv->gvt);
> > -	dev_priv->gvt = NULL;
> > +	kfree(i915->gvt);
> > +	i915->gvt = NULL;
> >   }
> >   /**
> >    * intel_gvt_init_device - initialize a GVT device
> > - * @dev_priv: drm i915 private data
> > + * @i915: drm i915 private data
> >    *
> >    * This function is called at the initialization stage, to initialize
> >    * necessary GVT components.
> > @@ -300,13 +300,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv)
> >    * Zero on success, negative error code if failed.
> >    *
> >    */
> > -int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> > +int intel_gvt_init_device(struct drm_i915_private *i915)
> >   {
> >   	struct intel_gvt *gvt;
> >   	struct intel_vgpu *vgpu;
> >   	int ret;
> > -	if (WARN_ON(dev_priv->gvt))
> > +	if (WARN_ON(i915->gvt))
> >   		return -EEXIST;
> >   	gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL);
> > @@ -319,7 +319,7 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> >   	spin_lock_init(&gvt->scheduler.mmio_context_lock);
> >   	mutex_init(&gvt->lock);
> >   	mutex_init(&gvt->sched_lock);
> > -	gvt->dev_priv = dev_priv;
> > +	gvt->i915 = i915;
> >   	init_device_info(gvt);
> > @@ -380,8 +380,8 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv)
> >   		gvt_err("debugfs registration failed, go on.\n");
> >   	gvt_dbg_core("gvt device initialization is done\n");
> > -	dev_priv->gvt = gvt;
> > -	intel_gvt_host.dev = &dev_priv->drm.pdev->dev;
> > +	i915->gvt = gvt;
> > +	intel_gvt_host.dev = &i915->drm.pdev->dev;
> >   	intel_gvt_host.initialized = true;
> >   	return 0;
> > diff --git a/drivers/gpu/drm/i915/gvt/gvt.h b/drivers/gpu/drm/i915/gvt/gvt.h
> > index b54f2bdc13a4..fbebe0ae4096 100644
> > --- a/drivers/gpu/drm/i915/gvt/gvt.h
> > +++ b/drivers/gpu/drm/i915/gvt/gvt.h
> > @@ -306,7 +306,7 @@ struct intel_gvt {
> >   	/* scheduler scope lock, protect gvt and vgpu schedule related data */
> >   	struct mutex sched_lock;
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	struct idr vgpu_idr;	/* vGPU IDR pool */
> >   	struct intel_gvt_device_info device_info;
> > @@ -373,12 +373,12 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt);
> >   #define HOST_FENCE 4
> >   /* Aperture/GM space definitions for GVT device */
> > -#define gvt_aperture_sz(gvt)	  (gvt->dev_priv->ggtt.mappable_end)
> > -#define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start)
> > +#define gvt_aperture_sz(gvt)	  (gvt->i915->ggtt.mappable_end)
> > +#define gvt_aperture_pa_base(gvt) (gvt->i915->ggtt.gmadr.start)
> > -#define gvt_ggtt_gm_sz(gvt)	  (gvt->dev_priv->ggtt.vm.total)
> > +#define gvt_ggtt_gm_sz(gvt)	  (gvt->i915->ggtt.vm.total)
> >   #define gvt_ggtt_sz(gvt) \
> > -	((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3)
> > +	((gvt->i915->ggtt.vm.total >> PAGE_SHIFT) << 3)
> >   #define gvt_hidden_sz(gvt)	  (gvt_ggtt_gm_sz(gvt) - gvt_aperture_sz(gvt))
> >   #define gvt_aperture_gmadr_base(gvt) (0)
> > @@ -390,7 +390,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt);
> >   #define gvt_hidden_gmadr_end(gvt) (gvt_hidden_gmadr_base(gvt) \
> >   				   + gvt_hidden_sz(gvt) - 1)
> > -#define gvt_fence_sz(gvt) (gvt->dev_priv->num_fence_regs)
> > +#define gvt_fence_sz(gvt) (gvt->i915->num_fence_regs)
> >   /* Aperture/GM space definitions for vGPU */
> >   #define vgpu_aperture_offset(vgpu)	((vgpu)->gm.low_gm_node.start)
> > @@ -582,14 +582,14 @@ enum {
> >   	GVT_FAILSAFE_GUEST_ERR,
> >   };
> > -static inline void mmio_hw_access_pre(struct drm_i915_private *dev_priv)
> > +static inline void mmio_hw_access_pre(struct drm_i915_private *i915)
> >   {
> > -	intel_runtime_pm_get(dev_priv);
> > +	intel_runtime_pm_get(i915);
> >   }
> > -static inline void mmio_hw_access_post(struct drm_i915_private *dev_priv)
> > +static inline void mmio_hw_access_post(struct drm_i915_private *i915)
> >   {
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   }
> >   /**
> > diff --git a/drivers/gpu/drm/i915/gvt/handlers.c b/drivers/gpu/drm/i915/gvt/handlers.c
> > index a6ade66349bd..9acb01b0b7c7 100644
> > --- a/drivers/gpu/drm/i915/gvt/handlers.c
> > +++ b/drivers/gpu/drm/i915/gvt/handlers.c
> > @@ -49,15 +49,15 @@
> >   unsigned long intel_gvt_get_device_type(struct intel_gvt *gvt)
> >   {
> > -	if (IS_BROADWELL(gvt->dev_priv))
> > +	if (IS_BROADWELL(gvt->i915))
> >   		return D_BDW;
> > -	else if (IS_SKYLAKE(gvt->dev_priv))
> > +	else if (IS_SKYLAKE(gvt->i915))
> >   		return D_SKL;
> > -	else if (IS_KABYLAKE(gvt->dev_priv))
> > +	else if (IS_KABYLAKE(gvt->i915))
> >   		return D_KBL;
> > -	else if (IS_BROXTON(gvt->dev_priv))
> > +	else if (IS_BROXTON(gvt->i915))
> >   		return D_BXT;
> > -	else if (IS_COFFEELAKE(gvt->dev_priv))
> > +	else if (IS_COFFEELAKE(gvt->i915))
> >   		return D_CFL;
> >   	return 0;
> > @@ -156,7 +156,7 @@ int intel_gvt_render_mmio_to_ring_id(struct intel_gvt *gvt,
> >   	struct intel_engine_cs *engine;
> >   	offset &= ~GENMASK(11, 0);
> > -	for_each_engine(engine, gvt->dev_priv, id) {
> > +	for_each_engine(engine, gvt->i915, id) {
> >   		if (engine->mmio_base == offset)
> >   			return id;
> >   	}
> > @@ -217,7 +217,7 @@ static int gamw_echo_dev_rw_ia_write(struct intel_vgpu *vgpu,
> >   {
> >   	u32 ips = (*(u32 *)p_data) & GAMW_ECO_ENABLE_64K_IPS_FIELD;
> > -	if (INTEL_GEN(vgpu->gvt->dev_priv) <= 10) {
> > +	if (INTEL_GEN(vgpu->gvt->i915) <= 10) {
> >   		if (ips == GAMW_ECO_ENABLE_64K_IPS_FIELD)
> >   			gvt_dbg_core("vgpu%d: ips enabled\n", vgpu->id);
> >   		else if (!ips)
> > @@ -253,7 +253,7 @@ static int fence_mmio_read(struct intel_vgpu *vgpu, unsigned int off,
> >   static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> >   		void *p_data, unsigned int bytes)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	unsigned int fence_num = offset_to_fence_num(off);
> >   	int ret;
> > @@ -262,10 +262,10 @@ static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off,
> >   		return ret;
> >   	write_vreg(vgpu, off, p_data, bytes);
> > -	mmio_hw_access_pre(dev_priv);
> > +	mmio_hw_access_pre(i915);
> >   	intel_vgpu_write_fence(vgpu, fence_num,
> >   			vgpu_vreg64(vgpu, fence_num_to_offset(fence_num)));
> > -	mmio_hw_access_post(dev_priv);
> > +	mmio_hw_access_post(i915);
> >   	return 0;
> >   }
> > @@ -283,7 +283,7 @@ static int mul_force_wake_write(struct intel_vgpu *vgpu,
> >   	old = vgpu_vreg(vgpu, offset);
> >   	new = CALC_MODE_MASK_REG(old, *(u32 *)p_data);
> > -	if (INTEL_GEN(vgpu->gvt->dev_priv)  >=  9) {
> > +	if (INTEL_GEN(vgpu->gvt->i915)  >=  9) {
> >   		switch (offset) {
> >   		case FORCEWAKE_RENDER_GEN9_REG:
> >   			ack_reg_offset = FORCEWAKE_ACK_RENDER_GEN9_REG;
> > @@ -341,7 +341,7 @@ static int gdrst_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   			gvt_dbg_mmio("vgpu%d: request VCS2 Reset\n", vgpu->id);
> >   			engine_mask |= BIT(VCS1);
> >   		}
> > -		engine_mask &= INTEL_INFO(vgpu->gvt->dev_priv)->engine_mask;
> > +		engine_mask &= INTEL_INFO(vgpu->gvt->i915)->engine_mask;
> >   	}
> >   	/* vgpu_lock already hold by emulate mmio r/w */
> > @@ -511,7 +511,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu,
> >   	u32 reg_nonpriv = *(u32 *)p_data;
> >   	int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset);
> >   	u32 ring_base;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int ret = -EINVAL;
> >   	if ((bytes != 4) || ((offset & (bytes - 1)) != 0) || ring_id < 0) {
> > @@ -520,7 +520,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu,
> >   		return ret;
> >   	}
> > -	ring_base = dev_priv->engine[ring_id]->mmio_base;
> > +	ring_base = i915->engine[ring_id]->mmio_base;
> >   	if (in_whitelist(reg_nonpriv) ||
> >   		reg_nonpriv == i915_mmio_reg_offset(RING_NOPID(ring_base))) {
> > @@ -751,7 +751,7 @@ static int south_chicken2_mmio_write(struct intel_vgpu *vgpu,
> >   static int pri_surf_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   		void *p_data, unsigned int bytes)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	u32 pipe = DSPSURF_TO_PIPE(offset);
> >   	int event = SKL_FLIP_EVENT(pipe, PLANE_PRIMARY);
> > @@ -792,7 +792,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu,
> >   			       unsigned int offset, void *p_data,
> >   			       unsigned int bytes)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	enum pipe pipe = REG_50080_TO_PIPE(offset);
> >   	enum plane_id plane = REG_50080_TO_PLANE(offset);
> >   	int event = SKL_FLIP_EVENT(pipe, plane);
> > @@ -816,7 +816,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu,
> >   static int trigger_aux_channel_interrupt(struct intel_vgpu *vgpu,
> >   		unsigned int reg)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	enum intel_gvt_event_type event;
> >   	if (reg == _DPA_AUX_CH_CTL)
> > @@ -916,11 +916,11 @@ static int dp_aux_ch_ctl_mmio_write(struct intel_vgpu *vgpu,
> >   	write_vreg(vgpu, offset, p_data, bytes);
> >   	data = vgpu_vreg(vgpu, offset);
> > -	if ((INTEL_GEN(vgpu->gvt->dev_priv) >= 9)
> > +	if ((INTEL_GEN(vgpu->gvt->i915) >= 9)
> >   		&& offset != _REG_SKL_DP_AUX_CH_CTL(port_index)) {
> >   		/* SKL DPB/C/D aux ctl register changed */
> >   		return 0;
> > -	} else if (IS_BROADWELL(vgpu->gvt->dev_priv) &&
> > +	} else if (IS_BROADWELL(vgpu->gvt->i915) &&
> >   		   offset != _REG_HSW_DP_AUX_CH_CTL(port_index)) {
> >   		/* write to the data registers */
> >   		return 0;
> > @@ -1236,8 +1236,8 @@ static int handle_g2v_notification(struct intel_vgpu *vgpu, int notification)
> >   static int send_display_ready_uevent(struct intel_vgpu *vgpu, int ready)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > -	struct kobject *kobj = &dev_priv->drm.primary->kdev->kobj;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > +	struct kobject *kobj = &i915->drm.primary->kdev->kobj;
> >   	char *env[3] = {NULL, NULL, NULL};
> >   	char vmid_str[20];
> >   	char display_ready_str[20];
> > @@ -1416,9 +1416,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   	switch (cmd) {
> >   	case GEN9_PCODE_READ_MEM_LATENCY:
> > -		if (IS_SKYLAKE(vgpu->gvt->dev_priv)
> > -			 || IS_KABYLAKE(vgpu->gvt->dev_priv)
> > -			 || IS_COFFEELAKE(vgpu->gvt->dev_priv)) {
> > +		if (IS_SKYLAKE(vgpu->gvt->i915)
> > +			 || IS_KABYLAKE(vgpu->gvt->i915)
> > +			 || IS_COFFEELAKE(vgpu->gvt->i915)) {
> >   			/**
> >   			 * "Read memory latency" command on gen9.
> >   			 * Below memory latency values are read
> > @@ -1428,7 +1428,7 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   				*data0 = 0x1e1a1100;
> >   			else
> >   				*data0 = 0x61514b3d;
> > -		} else if (IS_BROXTON(vgpu->gvt->dev_priv)) {
> > +		} else if (IS_BROXTON(vgpu->gvt->i915)) {
> >   			/**
> >   			 * "Read memory latency" command on gen9.
> >   			 * Below memory latency values are read
> > @@ -1441,9 +1441,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   		}
> >   		break;
> >   	case SKL_PCODE_CDCLK_CONTROL:
> > -		if (IS_SKYLAKE(vgpu->gvt->dev_priv)
> > -			 || IS_KABYLAKE(vgpu->gvt->dev_priv)
> > -			 || IS_COFFEELAKE(vgpu->gvt->dev_priv))
> > +		if (IS_SKYLAKE(vgpu->gvt->i915)
> > +			 || IS_KABYLAKE(vgpu->gvt->i915)
> > +			 || IS_COFFEELAKE(vgpu->gvt->i915))
> >   			*data0 = SKL_CDCLK_READY_FOR_CHANGE;
> >   		break;
> >   	case GEN6_PCODE_READ_RC6VIDS:
> > @@ -1496,7 +1496,7 @@ static int skl_power_well_ctl_write(struct intel_vgpu *vgpu,
> >   {
> >   	u32 v = *(u32 *)p_data;
> > -	if (IS_BROXTON(vgpu->gvt->dev_priv))
> > +	if (IS_BROXTON(vgpu->gvt->i915))
> >   		v &= (1 << 31) | (1 << 29);
> >   	else
> >   		v &= (1 << 31) | (1 << 29) | (1 << 9) |
> > @@ -1633,7 +1633,7 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu,
> >   		unsigned int offset, void *p_data, unsigned int bytes)
> >   {
> >   	struct intel_gvt *gvt = vgpu->gvt;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ring_id;
> >   	u32 ring_base;
> > @@ -1645,14 +1645,14 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu,
> >   	 * c. the offset is ring time stamp mmio
> >   	 */
> >   	if (ring_id >= 0)
> > -		ring_base = dev_priv->engine[ring_id]->mmio_base;
> > +		ring_base = i915->engine[ring_id]->mmio_base;
> >   	if (ring_id < 0 || vgpu  == gvt->scheduler.engine_owner[ring_id] ||
> >   	    offset == i915_mmio_reg_offset(RING_TIMESTAMP(ring_base)) ||
> >   	    offset == i915_mmio_reg_offset(RING_TIMESTAMP_UDW(ring_base))) {
> > -		mmio_hw_access_pre(dev_priv);
> > +		mmio_hw_access_pre(i915);
> >   		vgpu_vreg(vgpu, offset) = I915_READ(_MMIO(offset));
> > -		mmio_hw_access_post(dev_priv);
> > +		mmio_hw_access_post(i915);
> >   	}
> >   	return intel_vgpu_default_mmio_read(vgpu, offset, p_data, bytes);
> > @@ -1693,7 +1693,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   	int ret;
> >   	(*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(1);
> > -	if (IS_COFFEELAKE(vgpu->gvt->dev_priv))
> > +	if (IS_COFFEELAKE(vgpu->gvt->i915))
> >   		(*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(2);
> >   	write_vreg(vgpu, offset, p_data, bytes);
> > @@ -1702,7 +1702,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset,
> >   		return 0;
> >   	}
> > -	if (IS_COFFEELAKE(vgpu->gvt->dev_priv) &&
> > +	if (IS_COFFEELAKE(vgpu->gvt->i915) &&
> >   	    data & _MASKED_BIT_ENABLE(2)) {
> >   		enter_failsafe_mode(vgpu, GVT_FAILSAFE_UNSUPPORTED_GUEST);
> >   		return 0;
> > @@ -1834,7 +1834,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu,
> >   	MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \
> >   	MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \
> >   	MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \
> > -	if (HAS_ENGINE(dev_priv, VCS1)) \
> > +	if (HAS_ENGINE(i915, VCS1)) \
> >   		MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \
> >   } while (0)
> > @@ -1855,7 +1855,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu,
> >   static int init_generic_mmio_info(struct intel_gvt *gvt)
> >   {
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ret;
> >   	MMIO_RING_DFH(RING_IMR, D_ALL, F_CMD_ACCESS, NULL,
> > @@ -2670,7 +2670,7 @@ static int init_generic_mmio_info(struct intel_gvt *gvt)
> >   static int init_broadwell_mmio_info(struct intel_gvt *gvt)
> >   {
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ret;
> >   	MMIO_DH(GEN8_GT_IMR(0), D_BDW_PLUS, NULL, intel_vgpu_reg_imr_handler);
> > @@ -2859,7 +2859,7 @@ static int init_broadwell_mmio_info(struct intel_gvt *gvt)
> >   static int init_skl_mmio_info(struct intel_gvt *gvt)
> >   {
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ret;
> >   	MMIO_DH(FORCEWAKE_RENDER_GEN9, D_SKL_PLUS, NULL, mul_force_wake_write);
> > @@ -3108,7 +3108,7 @@ static int init_skl_mmio_info(struct intel_gvt *gvt)
> >   static int init_bxt_mmio_info(struct intel_gvt *gvt)
> >   {
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int ret;
> >   	MMIO_F(_MMIO(0x80000), 0x3000, 0, 0, 0, D_BXT, NULL, NULL);
> > @@ -3344,7 +3344,7 @@ static struct gvt_mmio_block mmio_blocks[] = {
> >   int intel_gvt_setup_mmio_info(struct intel_gvt *gvt)
> >   {
> >   	struct intel_gvt_device_info *info = &gvt->device_info;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	int size = info->mmio_size / 4 * sizeof(*gvt->mmio.mmio_attribute);
> >   	int ret;
> > @@ -3356,20 +3356,20 @@ int intel_gvt_setup_mmio_info(struct intel_gvt *gvt)
> >   	if (ret)
> >   		goto err;
> > -	if (IS_BROADWELL(dev_priv)) {
> > +	if (IS_BROADWELL(i915)) {
> >   		ret = init_broadwell_mmio_info(gvt);
> >   		if (ret)
> >   			goto err;
> > -	} else if (IS_SKYLAKE(dev_priv)
> > -		|| IS_KABYLAKE(dev_priv)
> > -		|| IS_COFFEELAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)
> > +		|| IS_KABYLAKE(i915)
> > +		|| IS_COFFEELAKE(i915)) {
> >   		ret = init_broadwell_mmio_info(gvt);
> >   		if (ret)
> >   			goto err;
> >   		ret = init_skl_mmio_info(gvt);
> >   		if (ret)
> >   			goto err;
> > -	} else if (IS_BROXTON(dev_priv)) {
> > +	} else if (IS_BROXTON(i915)) {
> >   		ret = init_broadwell_mmio_info(gvt);
> >   		if (ret)
> >   			goto err;
> > diff --git a/drivers/gpu/drm/i915/gvt/interrupt.c b/drivers/gpu/drm/i915/gvt/interrupt.c
> > index 951681813230..37e2e72bfacc 100644
> > --- a/drivers/gpu/drm/i915/gvt/interrupt.c
> > +++ b/drivers/gpu/drm/i915/gvt/interrupt.c
> > @@ -536,7 +536,7 @@ static void gen8_init_irq(
> >   	SET_BIT_INFO(irq, 4, VCS_MI_FLUSH_DW, INTEL_GVT_IRQ_INFO_GT1);
> >   	SET_BIT_INFO(irq, 8, VCS_AS_CONTEXT_SWITCH, INTEL_GVT_IRQ_INFO_GT1);
> > -	if (HAS_ENGINE(gvt->dev_priv, VCS1)) {
> > +	if (HAS_ENGINE(gvt->i915, VCS1)) {
> >   		SET_BIT_INFO(irq, 16, VCS2_MI_USER_INTERRUPT,
> >   			INTEL_GVT_IRQ_INFO_GT1);
> >   		SET_BIT_INFO(irq, 20, VCS2_MI_FLUSH_DW,
> > @@ -568,7 +568,7 @@ static void gen8_init_irq(
> >   	SET_BIT_INFO(irq, 22, DP_C_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH);
> >   	SET_BIT_INFO(irq, 23, DP_D_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH);
> > -	if (IS_BROADWELL(gvt->dev_priv)) {
> > +	if (IS_BROADWELL(gvt->i915)) {
> >   		SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_PCH);
> >   		SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_PCH);
> >   		SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_PCH);
> > @@ -581,7 +581,7 @@ static void gen8_init_irq(
> >   		SET_BIT_INFO(irq, 4, PRIMARY_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C);
> >   		SET_BIT_INFO(irq, 5, SPRITE_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C);
> > -	} else if (INTEL_GEN(gvt->dev_priv) >= 9) {
> > +	} else if (INTEL_GEN(gvt->i915) >= 9) {
> >   		SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_DE_PORT);
> >   		SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_DE_PORT);
> >   		SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_DE_PORT);
> > diff --git a/drivers/gpu/drm/i915/gvt/kvmgt.c b/drivers/gpu/drm/i915/gvt/kvmgt.c
> > index 144301b778df..615a7959b762 100644
> > --- a/drivers/gpu/drm/i915/gvt/kvmgt.c
> > +++ b/drivers/gpu/drm/i915/gvt/kvmgt.c
> > @@ -187,7 +187,7 @@ static int gvt_pin_guest_page(struct intel_vgpu *vgpu, unsigned long gfn,
> >   static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn,
> >   		dma_addr_t *dma_addr, unsigned long size)
> >   {
> > -	struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> >   	struct page *page = NULL;
> >   	int ret;
> > @@ -210,7 +210,7 @@ static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn,
> >   static void gvt_dma_unmap_page(struct intel_vgpu *vgpu, unsigned long gfn,
> >   		dma_addr_t dma_addr, unsigned long size)
> >   {
> > -	struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev;
> > +	struct device *dev = &vgpu->gvt->i915->drm.pdev->dev;
> >   	dma_unmap_page(dev, dma_addr, size, PCI_DMA_BIDIRECTIONAL);
> >   	gvt_unpin_guest_page(vgpu, gfn, size);
> > @@ -913,7 +913,7 @@ static int intel_vgpu_aperture_rw(struct intel_vgpu *vgpu, u64 off,
> >   		return -EINVAL;
> >   	}
> > -	aperture_va = io_mapping_map_wc(&vgpu->gvt->dev_priv->ggtt.iomap,
> > +	aperture_va = io_mapping_map_wc(&vgpu->gvt->i915->ggtt.iomap,
> >   					ALIGN_DOWN(off, PAGE_SIZE),
> >   					count + offset_in_page(off));
> >   	if (!aperture_va)
> > diff --git a/drivers/gpu/drm/i915/gvt/mmio.c b/drivers/gpu/drm/i915/gvt/mmio.c
> > index a55178884d67..572b8328af61 100644
> > --- a/drivers/gpu/drm/i915/gvt/mmio.c
> > +++ b/drivers/gpu/drm/i915/gvt/mmio.c
> > @@ -245,7 +245,7 @@ void intel_vgpu_reset_mmio(struct intel_vgpu *vgpu, bool dmlr)
> >   		/* set the bit 0:2(Core C-State ) to C0 */
> >   		vgpu_vreg_t(vgpu, GEN6_GT_CORE_STATUS) = 0;
> > -		if (IS_BROXTON(vgpu->gvt->dev_priv)) {
> > +		if (IS_BROXTON(vgpu->gvt->i915)) {
> >   			vgpu_vreg_t(vgpu, BXT_P_CR_GT_DISP_PWRON) &=
> >   				    ~(BIT(0) | BIT(1));
> >   			vgpu_vreg_t(vgpu, BXT_PORT_CL1CM_DW0(DPIO_PHY0)) &=
> > diff --git a/drivers/gpu/drm/i915/gvt/mmio_context.c b/drivers/gpu/drm/i915/gvt/mmio_context.c
> > index 2998999e8568..44fd6a0a698c 100644
> > --- a/drivers/gpu/drm/i915/gvt/mmio_context.c
> > +++ b/drivers/gpu/drm/i915/gvt/mmio_context.c
> > @@ -148,7 +148,7 @@ static struct {
> >   	u32 l3cc_table[GEN9_MOCS_SIZE / 2];
> >   } gen9_render_mocs;
> > -static void load_render_mocs(struct drm_i915_private *dev_priv)
> > +static void load_render_mocs(struct drm_i915_private *i915)
> >   {
> >   	i915_reg_t offset;
> >   	u32 regs[] = {
> > @@ -161,7 +161,7 @@ static void load_render_mocs(struct drm_i915_private *dev_priv)
> >   	int ring_id, i;
> >   	for (ring_id = 0; ring_id < ARRAY_SIZE(regs); ring_id++) {
> > -		if (!HAS_ENGINE(dev_priv, ring_id))
> > +		if (!HAS_ENGINE(i915, ring_id))
> >   			continue;
> >   		offset.reg = regs[ring_id];
> >   		for (i = 0; i < GEN9_MOCS_SIZE; i++) {
> > @@ -329,8 +329,8 @@ int intel_vgpu_restore_inhibit_context(struct intel_vgpu *vgpu,
> >   static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> >   	enum forcewake_domains fw;
> >   	i915_reg_t reg;
> > @@ -357,7 +357,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> >   	 */
> >   	fw = intel_uncore_forcewake_for_reg(uncore, reg,
> >   					    FW_REG_READ | FW_REG_WRITE);
> > -	if (ring_id == RCS0 && INTEL_GEN(dev_priv) >= 9)
> > +	if (ring_id == RCS0 && INTEL_GEN(i915) >= 9)
> >   		fw |= FORCEWAKE_RENDER;
> >   	intel_uncore_forcewake_get(uncore, fw);
> > @@ -377,7 +377,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id)
> >   static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
> >   			int ring_id)
> >   {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	i915_reg_t offset, l3_offset;
> >   	u32 old_v, new_v;
> > @@ -390,15 +390,15 @@ static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next,
> >   	};
> >   	int i;
> > -	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> > +	i915 = pre ? pre->gvt->i915 : next->gvt->i915;
> >   	if (WARN_ON(ring_id >= ARRAY_SIZE(regs)))
> >   		return;
> > -	if (ring_id == RCS0 && IS_GEN(dev_priv, 9))
> > +	if (ring_id == RCS0 && IS_GEN(i915, 9))
> >   		return;
> >   	if (!pre && !gen9_render_mocs.initialized)
> > -		load_render_mocs(dev_priv);
> > +		load_render_mocs(i915);
> >   	offset.reg = regs[ring_id];
> >   	for (i = 0; i < GEN9_MOCS_SIZE; i++) {
> > @@ -454,16 +454,16 @@ static void switch_mmio(struct intel_vgpu *pre,
> >   			struct intel_vgpu *next,
> >   			int ring_id)
> >   {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	struct intel_vgpu_submission *s;
> >   	struct engine_mmio *mmio;
> >   	u32 old_v, new_v;
> > -	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	i915 = pre ? pre->gvt->i915 : next->gvt->i915;
> > +	if (INTEL_GEN(i915) >= 9)
> >   		switch_mocs(pre, next, ring_id);
> > -	for (mmio = dev_priv->gvt->engine_mmio_list.mmio;
> > +	for (mmio = i915->gvt->engine_mmio_list.mmio;
> >   	     i915_mmio_reg_valid(mmio->reg); mmio++) {
> >   		if (mmio->ring_id != ring_id)
> >   			continue;
> > @@ -472,7 +472,7 @@ static void switch_mmio(struct intel_vgpu *pre,
> >   		 * state image on gen9, it's initialized by lri command and
> >   		 * save or restore with context together.
> >   		 */
> > -		if (IS_GEN(dev_priv, 9) && mmio->in_context)
> > +		if (IS_GEN(i915, 9) && mmio->in_context)
> >   			continue;
> >   		// save
> > @@ -536,7 +536,7 @@ static void switch_mmio(struct intel_vgpu *pre,
> >   void intel_gvt_switch_mmio(struct intel_vgpu *pre,
> >   			   struct intel_vgpu *next, int ring_id)
> >   {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	if (WARN_ON(!pre && !next))
> >   		return;
> > @@ -544,16 +544,16 @@ void intel_gvt_switch_mmio(struct intel_vgpu *pre,
> >   	gvt_dbg_render("switch ring %d from %s to %s\n", ring_id,
> >   		       pre ? "vGPU" : "host", next ? "vGPU" : "HOST");
> > -	dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv;
> > +	i915 = pre ? pre->gvt->i915 : next->gvt->i915;
> >   	/**
> >   	 * We are using raw mmio access wrapper to improve the
> >   	 * performace for batch mmio read/write, so we need
> >   	 * handle forcewake mannually.
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	switch_mmio(pre, next, ring_id);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> >   /**
> > @@ -565,7 +565,7 @@ void intel_gvt_init_engine_mmio_context(struct intel_gvt *gvt)
> >   {
> >   	struct engine_mmio *mmio;
> > -	if (INTEL_GEN(gvt->dev_priv) >= 9)
> > +	if (INTEL_GEN(gvt->i915) >= 9)
> >   		gvt->engine_mmio_list.mmio = gen9_engine_mmio_list;
> >   	else
> >   		gvt->engine_mmio_list.mmio = gen8_engine_mmio_list;
> > diff --git a/drivers/gpu/drm/i915/gvt/sched_policy.c b/drivers/gpu/drm/i915/gvt/sched_policy.c
> > index 1c763a27a412..64a5d4688c89 100644
> > --- a/drivers/gpu/drm/i915/gvt/sched_policy.c
> > +++ b/drivers/gpu/drm/i915/gvt/sched_policy.c
> > @@ -39,7 +39,7 @@ static bool vgpu_has_pending_workload(struct intel_vgpu *vgpu)
> >   	enum intel_engine_id i;
> >   	struct intel_engine_cs *engine;
> > -	for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> > +	for_each_engine(engine, vgpu->gvt->i915, i) {
> >   		if (!list_empty(workload_q_head(vgpu, i)))
> >   			return true;
> >   	}
> > @@ -152,7 +152,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt)
> >   	scheduler->need_reschedule = true;
> >   	/* still have uncompleted workload? */
> > -	for_each_engine(engine, gvt->dev_priv, i) {
> > +	for_each_engine(engine, gvt->i915, i) {
> >   		if (scheduler->current_workload[i])
> >   			return;
> >   	}
> > @@ -169,7 +169,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt)
> >   	scheduler->need_reschedule = false;
> >   	/* wake up workload dispatch thread */
> > -	for_each_engine(engine, gvt->dev_priv, i)
> > +	for_each_engine(engine, gvt->i915, i)
> >   		wake_up(&scheduler->waitq[i]);
> >   }
> > @@ -446,7 +446,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> >   		&vgpu->gvt->scheduler;
> >   	int ring_id;
> >   	struct vgpu_sched_data *vgpu_data = vgpu->sched_data;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	if (!vgpu_data->active)
> >   		return;
> > @@ -465,7 +465,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> >   		scheduler->current_vgpu = NULL;
> >   	}
> > -	intel_runtime_pm_get(dev_priv);
> > +	intel_runtime_pm_get(i915);
> >   	spin_lock_bh(&scheduler->mmio_context_lock);
> >   	for (ring_id = 0; ring_id < I915_NUM_ENGINES; ring_id++) {
> >   		if (scheduler->engine_owner[ring_id] == vgpu) {
> > @@ -474,6 +474,6 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu)
> >   		}
> >   	}
> >   	spin_unlock_bh(&scheduler->mmio_context_lock);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   	mutex_unlock(&vgpu->gvt->sched_lock);
> >   }
> > diff --git a/drivers/gpu/drm/i915/gvt/scheduler.c b/drivers/gpu/drm/i915/gvt/scheduler.c
> > index e301efb18d45..ec7299d8a738 100644
> > --- a/drivers/gpu/drm/i915/gvt/scheduler.c
> > +++ b/drivers/gpu/drm/i915/gvt/scheduler.c
> > @@ -83,9 +83,9 @@ static void update_shadow_pdps(struct intel_vgpu_workload *workload)
> >   static void sr_oa_regs(struct intel_vgpu_workload *workload,
> >   		u32 *reg_state, bool save)
> >   {
> > -	struct drm_i915_private *dev_priv = workload->vgpu->gvt->dev_priv;
> > -	u32 ctx_oactxctrl = dev_priv->perf.oa.ctx_oactxctrl_offset;
> > -	u32 ctx_flexeu0 = dev_priv->perf.oa.ctx_flexeu0_offset;
> > +	struct drm_i915_private *i915 = workload->vgpu->gvt->i915;
> > +	u32 ctx_oactxctrl = i915->perf.oa.ctx_oactxctrl_offset;
> > +	u32 ctx_flexeu0 = i915->perf.oa.ctx_flexeu0_offset;
> >   	int i = 0;
> >   	u32 flex_mmio[] = {
> >   		i915_mmio_reg_offset(EU_PERF_CNTL0),
> > @@ -177,11 +177,11 @@ static int populate_shadow_context(struct intel_vgpu_workload *workload)
> >   	gvt_dbg_sched("ring id %d workload lrca %x", ring_id,
> >   			workload->ctx_desc.lrca);
> > -	context_page_num = gvt->dev_priv->engine[ring_id]->context_size;
> > +	context_page_num = gvt->i915->engine[ring_id]->context_size;
> >   	context_page_num = context_page_num >> PAGE_SHIFT;
> > -	if (IS_BROADWELL(gvt->dev_priv) && ring_id == RCS0)
> > +	if (IS_BROADWELL(gvt->i915) && ring_id == RCS0)
> >   		context_page_num = 19;
> >   	i = 2;
> > @@ -211,8 +211,8 @@ static inline bool is_gvt_request(struct i915_request *req)
> >   static void save_ring_hw_state(struct intel_vgpu *vgpu, int ring_id)
> >   {
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > -	u32 ring_base = dev_priv->engine[ring_id]->mmio_base;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> > +	u32 ring_base = i915->engine[ring_id]->mmio_base;
> >   	i915_reg_t reg;
> >   	reg = RING_INSTDONE(ring_base);
> > @@ -391,10 +391,10 @@ intel_gvt_workload_req_alloc(struct intel_vgpu_workload *workload)
> >   {
> >   	struct intel_vgpu *vgpu = workload->vgpu;
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct i915_request *rq;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	if (workload->req)
> >   		return 0;
> > @@ -421,10 +421,10 @@ int intel_gvt_scan_and_shadow_workload(struct intel_vgpu_workload *workload)
> >   {
> >   	struct intel_vgpu *vgpu = workload->vgpu;
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	int ret;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	if (workload->shadow)
> >   		return 0;
> > @@ -577,7 +577,7 @@ static int prepare_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx)
> >   static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> >   {
> >   	struct intel_vgpu *vgpu = workload->vgpu;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_vgpu_shadow_bb *bb, *pos;
> >   	if (list_empty(&workload->shadow_bb))
> > @@ -586,7 +586,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> >   	bb = list_first_entry(&workload->shadow_bb,
> >   			struct intel_vgpu_shadow_bb, list);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	list_for_each_entry_safe(bb, pos, &workload->shadow_bb, list) {
> >   		if (bb->obj) {
> > @@ -606,7 +606,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload)
> >   		kfree(bb);
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   }
> >   static int prepare_workload(struct intel_vgpu_workload *workload)
> > @@ -671,7 +671,7 @@ static int prepare_workload(struct intel_vgpu_workload *workload)
> >   static int dispatch_workload(struct intel_vgpu_workload *workload)
> >   {
> >   	struct intel_vgpu *vgpu = workload->vgpu;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> >   	struct i915_request *rq;
> >   	int ring_id = workload->ring_id;
> > @@ -681,7 +681,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload)
> >   		ring_id, workload);
> >   	mutex_lock(&vgpu->vgpu_lock);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	ret = set_context_ppgtt_from_shadow(workload,
> >   					    s->shadow[ring_id]->gem_context);
> > @@ -723,7 +723,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload)
> >   err_req:
> >   	if (ret)
> >   		workload->status = ret;
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	mutex_unlock(&vgpu->vgpu_lock);
> >   	return ret;
> >   }
> > @@ -796,7 +796,7 @@ static void update_guest_context(struct intel_vgpu_workload *workload)
> >   	void *src;
> >   	unsigned long context_gpa, context_page_num;
> >   	int i;
> > -	struct drm_i915_private *dev_priv = gvt->dev_priv;
> > +	struct drm_i915_private *i915 = gvt->i915;
> >   	u32 ring_base;
> >   	u32 head, tail;
> >   	u16 wrap_count;
> > @@ -817,14 +817,14 @@ static void update_guest_context(struct intel_vgpu_workload *workload)
> >   	head = (wrap_count << RB_HEAD_WRAP_CNT_OFF) | tail;
> > -	ring_base = dev_priv->engine[workload->ring_id]->mmio_base;
> > +	ring_base = i915->engine[workload->ring_id]->mmio_base;
> >   	vgpu_vreg_t(vgpu, RING_TAIL(ring_base)) = tail;
> >   	vgpu_vreg_t(vgpu, RING_HEAD(ring_base)) = head;
> >   	context_page_num = rq->engine->context_size;
> >   	context_page_num = context_page_num >> PAGE_SHIFT;
> > -	if (IS_BROADWELL(gvt->dev_priv) && rq->engine->id == RCS0)
> > +	if (IS_BROADWELL(gvt->i915) && rq->engine->id == RCS0)
> >   		context_page_num = 19;
> >   	i = 2;
> > @@ -875,13 +875,13 @@ void intel_vgpu_clean_workloads(struct intel_vgpu *vgpu,
> >   				intel_engine_mask_t engine_mask)
> >   {
> >   	struct intel_vgpu_submission *s = &vgpu->submission;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	struct intel_engine_cs *engine;
> >   	struct intel_vgpu_workload *pos, *n;
> >   	intel_engine_mask_t tmp;
> >   	/* free the unsubmited workloads in the queues. */
> > -	for_each_engine_masked(engine, dev_priv, engine_mask, tmp) {
> > +	for_each_engine_masked(engine, i915, engine_mask, tmp) {
> >   		list_for_each_entry_safe(pos, n,
> >   			&s->workload_q_head[engine->id], list) {
> >   			list_del_init(&pos->list);
> > @@ -986,7 +986,7 @@ static int workload_thread(void *priv)
> >   	struct intel_vgpu_workload *workload = NULL;
> >   	struct intel_vgpu *vgpu = NULL;
> >   	int ret;
> > -	bool need_force_wake = (INTEL_GEN(gvt->dev_priv) >= 9);
> > +	bool need_force_wake = (INTEL_GEN(gvt->i915) >= 9);
> >   	DEFINE_WAIT_FUNC(wait, woken_wake_function);
> >   	kfree(p);
> > @@ -1015,7 +1015,7 @@ static int workload_thread(void *priv)
> >   				workload->ring_id, workload);
> >   		if (need_force_wake)
> > -			intel_uncore_forcewake_get(&gvt->dev_priv->uncore,
> > +			intel_uncore_forcewake_get(&gvt->i915->uncore,
> >   					FORCEWAKE_ALL);
> >   		ret = dispatch_workload(workload);
> > @@ -1037,7 +1037,7 @@ static int workload_thread(void *priv)
> >   		complete_current_workload(gvt, ring_id);
> >   		if (need_force_wake)
> > -			intel_uncore_forcewake_put(&gvt->dev_priv->uncore,
> > +			intel_uncore_forcewake_put(&gvt->i915->uncore,
> >   					FORCEWAKE_ALL);
> >   		if (ret && (vgpu_is_vm_unhealthy(ret)))
> > @@ -1068,7 +1068,7 @@ void intel_gvt_clean_workload_scheduler(struct intel_gvt *gvt)
> >   	gvt_dbg_core("clean workload scheduler\n");
> > -	for_each_engine(engine, gvt->dev_priv, i) {
> > +	for_each_engine(engine, gvt->i915, i) {
> >   		atomic_notifier_chain_unregister(
> >   					&engine->context_status_notifier,
> >   					&gvt->shadow_ctx_notifier_block[i]);
> > @@ -1088,7 +1088,7 @@ int intel_gvt_init_workload_scheduler(struct intel_gvt *gvt)
> >   	init_waitqueue_head(&scheduler->workload_complete_wq);
> > -	for_each_engine(engine, gvt->dev_priv, i) {
> > +	for_each_engine(engine, gvt->i915, i) {
> >   		init_waitqueue_head(&scheduler->waitq[i]);
> >   		param = kzalloc(sizeof(*param), GFP_KERNEL);
> > @@ -1152,7 +1152,7 @@ void intel_vgpu_clean_submission(struct intel_vgpu *vgpu)
> >   	intel_vgpu_select_submission_ops(vgpu, ALL_ENGINES, 0);
> >   	i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(s->shadow[0]->gem_context->vm));
> > -	for_each_engine(engine, vgpu->gvt->dev_priv, id)
> > +	for_each_engine(engine, vgpu->gvt->i915, id)
> >   		intel_context_unpin(s->shadow[id]);
> >   	kmem_cache_destroy(s->workloads);
> > @@ -1212,13 +1212,13 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu)
> >   	enum intel_engine_id i;
> >   	int ret;
> > -	ctx = i915_gem_context_create_gvt(&vgpu->gvt->dev_priv->drm);
> > +	ctx = i915_gem_context_create_gvt(&vgpu->gvt->i915->drm);
> >   	if (IS_ERR(ctx))
> >   		return PTR_ERR(ctx);
> >   	i915_context_ppgtt_root_save(s, i915_vm_to_ppgtt(ctx->vm));
> > -	for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> > +	for_each_engine(engine, vgpu->gvt->i915, i) {
> >   		struct intel_context *ce;
> >   		INIT_LIST_HEAD(&s->workload_q_head[i]);
> > @@ -1260,7 +1260,7 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu)
> >   out_shadow_ctx:
> >   	i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(ctx->vm));
> > -	for_each_engine(engine, vgpu->gvt->dev_priv, i) {
> > +	for_each_engine(engine, vgpu->gvt->i915, i) {
> >   		if (IS_ERR(s->shadow[i]))
> >   			break;
> > @@ -1439,7 +1439,7 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id,
> >   	struct list_head *q = workload_q_head(vgpu, ring_id);
> >   	struct intel_vgpu_workload *last_workload = get_last_workload(q);
> >   	struct intel_vgpu_workload *workload = NULL;
> > -	struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv;
> > +	struct drm_i915_private *i915 = vgpu->gvt->i915;
> >   	u64 ring_context_gpa;
> >   	u32 head, tail, start, ctl, ctx_ctl, per_ctx, indirect_ctx;
> >   	u32 guest_head;
> > @@ -1526,11 +1526,11 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id,
> >   	 * as there is only one pre-allocated buf-obj for shadow.
> >   	 */
> >   	if (list_empty(workload_q_head(vgpu, ring_id))) {
> > -		intel_runtime_pm_get(dev_priv);
> > -		mutex_lock(&dev_priv->drm.struct_mutex);
> > +		intel_runtime_pm_get(i915);
> > +		mutex_lock(&i915->drm.struct_mutex);
> >   		ret = intel_gvt_scan_and_shadow_workload(workload);
> > -		mutex_unlock(&dev_priv->drm.struct_mutex);
> > -		intel_runtime_pm_put_unchecked(dev_priv);
> > +		mutex_unlock(&i915->drm.struct_mutex);
> > +		intel_runtime_pm_put_unchecked(i915);
> >   	}
> >   	if (ret) {
> > diff --git a/drivers/gpu/drm/i915/gvt/vgpu.c b/drivers/gpu/drm/i915/gvt/vgpu.c
> > index 44ce3c2b9ac1..a311737c9633 100644
> > --- a/drivers/gpu/drm/i915/gvt/vgpu.c
> > +++ b/drivers/gpu/drm/i915/gvt/vgpu.c
> > @@ -148,10 +148,10 @@ int intel_gvt_init_vgpu_types(struct intel_gvt *gvt)
> >   		gvt->types[i].avail_instance = min(low_avail / vgpu_types[i].low_mm,
> >   						   high_avail / vgpu_types[i].high_mm);
> > -		if (IS_GEN(gvt->dev_priv, 8))
> > +		if (IS_GEN(gvt->i915, 8))
> >   			sprintf(gvt->types[i].name, "GVTg_V4_%s",
> >   						vgpu_types[i].name);
> > -		else if (IS_GEN(gvt->dev_priv, 9))
> > +		else if (IS_GEN(gvt->i915, 9))
> >   			sprintf(gvt->types[i].name, "GVTg_V5_%s",
> >   						vgpu_types[i].name);
> > @@ -429,7 +429,7 @@ static struct intel_vgpu *__intel_gvt_create_vgpu(struct intel_gvt *gvt,
> >   		goto out_clean_sched_policy;
> >   	/*TODO: add more platforms support */
> > -	if (IS_SKYLAKE(gvt->dev_priv) || IS_KABYLAKE(gvt->dev_priv))
> > +	if (IS_SKYLAKE(gvt->i915) || IS_KABYLAKE(gvt->i915))
> >   		ret = intel_gvt_hypervisor_set_edid(vgpu, PORT_D);
> >   	if (ret)
> >   		goto out_clean_sched_policy;
> > diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c
> > index a28bcd2d7c09..e9d076167b24 100644
> > --- a/drivers/gpu/drm/i915/i915_cmd_parser.c
> > +++ b/drivers/gpu/drm/i915/i915_cmd_parser.c
> > @@ -1342,21 +1342,21 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
> >   /**
> >    * i915_cmd_parser_get_version() - get the cmd parser version number
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    *
> >    * The cmd parser maintains a simple increasing integer version number suitable
> >    * for passing to userspace clients to determine what operations are permitted.
> >    *
> >    * Return: the current version number of the cmd parser
> >    */
> > -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv)
> > +int i915_cmd_parser_get_version(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	bool active = false;
> >   	/* If the command parser is not enabled, report 0 - unsupported */
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		if (intel_engine_needs_cmd_parser(engine)) {
> >   			active = true;
> >   			break;
> > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
> > index 026fb46020f6..fec01858685d 100644
> > --- a/drivers/gpu/drm/i915/i915_debugfs.c
> > +++ b/drivers/gpu/drm/i915/i915_debugfs.c
> > @@ -55,17 +55,17 @@ static inline struct drm_i915_private *node_to_i915(struct drm_info_node *node)
> >   static int i915_capabilities(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	const struct intel_device_info *info = INTEL_INFO(dev_priv);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	const struct intel_device_info *info = INTEL_INFO(i915);
> >   	struct drm_printer p = drm_seq_file_printer(m);
> > -	seq_printf(m, "gen: %d\n", INTEL_GEN(dev_priv));
> > +	seq_printf(m, "gen: %d\n", INTEL_GEN(i915));
> >   	seq_printf(m, "platform: %s\n", intel_platform_name(info->platform));
> > -	seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(dev_priv));
> > +	seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(i915));
> >   	intel_device_info_dump_flags(info, &p);
> > -	intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p);
> > -	intel_driver_caps_print(&dev_priv->caps, &p);
> > +	intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p);
> > +	intel_driver_caps_print(&i915->caps, &p);
> >   	kernel_param_lock(THIS_MODULE);
> >   	i915_params_dump(&i915_modparams, &p);
> > @@ -137,7 +137,7 @@ stringify_page_sizes(unsigned int page_sizes, char *buf, size_t len)
> >   static void
> >   describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct intel_engine_cs *engine;
> >   	struct i915_vma *vma;
> >   	unsigned int frontbuffer_bits;
> > @@ -155,7 +155,7 @@ describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj)
> >   		   obj->base.size / 1024,
> >   		   obj->read_domains,
> >   		   obj->write_domain,
> > -		   i915_cache_level_str(dev_priv, obj->cache_level),
> > +		   i915_cache_level_str(i915, obj->cache_level),
> >   		   obj->mm.dirty ? " dirty" : "",
> >   		   obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : "");
> >   	if (obj->base.name)
> > @@ -296,7 +296,7 @@ static int per_file_stats(int id, void *ptr, void *data)
> >   } while (0)
> >   static void print_batch_pool_stats(struct seq_file *m,
> > -				   struct drm_i915_private *dev_priv)
> > +				   struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_gem_object *obj;
> >   	struct intel_engine_cs *engine;
> > @@ -304,7 +304,7 @@ static void print_batch_pool_stats(struct seq_file *m,
> >   	enum intel_engine_id id;
> >   	int j;
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) {
> >   			list_for_each_entry(obj,
> >   					    &engine->batch_pool.cache_list[j],
> > @@ -382,8 +382,8 @@ static int i915_gem_object_info(struct seq_file *m, void *data)
> >   static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_i915_gem_object *obj;
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -394,7 +394,7 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
> >   	if (ret)
> >   		return ret;
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) {
> >   			int count;
> > @@ -427,15 +427,15 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data)
> >   static void gen8_display_interrupt_info(struct seq_file *m)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	int pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		enum intel_display_power_domain power_domain;
> >   		intel_wakeref_t wakeref;
> >   		power_domain = POWER_DOMAIN_PIPE(pipe);
> > -		wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +		wakeref = intel_display_power_get_if_enabled(i915,
> >   							     power_domain);
> >   		if (!wakeref) {
> >   			seq_printf(m, "Pipe %c power disabled\n",
> > @@ -452,7 +452,7 @@ static void gen8_display_interrupt_info(struct seq_file *m)
> >   			   pipe_name(pipe),
> >   			   I915_READ(GEN8_DE_PIPE_IER(pipe)));
> > -		intel_display_power_put(dev_priv, power_domain, wakeref);
> > +		intel_display_power_put(i915, power_domain, wakeref);
> >   	}
> >   	seq_printf(m, "Display Engine port interrupt mask:\t%08x\n",
> > @@ -479,15 +479,15 @@ static void gen8_display_interrupt_info(struct seq_file *m)
> >   static int i915_interrupt_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	intel_wakeref_t wakeref;
> >   	int i, pipe;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		intel_wakeref_t pref;
> >   		seq_printf(m, "Master Interrupt Control:\t%08x\n",
> > @@ -501,11 +501,11 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			   I915_READ(VLV_IIR_RW));
> >   		seq_printf(m, "Display IMR:\t%08x\n",
> >   			   I915_READ(VLV_IMR));
> > -		for_each_pipe(dev_priv, pipe) {
> > +		for_each_pipe(i915, pipe) {
> >   			enum intel_display_power_domain power_domain;
> >   			power_domain = POWER_DOMAIN_PIPE(pipe);
> > -			pref = intel_display_power_get_if_enabled(dev_priv,
> > +			pref = intel_display_power_get_if_enabled(i915,
> >   								  power_domain);
> >   			if (!pref) {
> >   				seq_printf(m, "Pipe %c power disabled\n",
> > @@ -517,17 +517,17 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   				   pipe_name(pipe),
> >   				   I915_READ(PIPESTAT(pipe)));
> > -			intel_display_power_put(dev_priv, power_domain, pref);
> > +			intel_display_power_put(i915, power_domain, pref);
> >   		}
> > -		pref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> > +		pref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
> >   		seq_printf(m, "Port hotplug:\t%08x\n",
> >   			   I915_READ(PORT_HOTPLUG_EN));
> >   		seq_printf(m, "DPFLIPSTAT:\t%08x\n",
> >   			   I915_READ(VLV_DPFLIPSTAT));
> >   		seq_printf(m, "DPINVGTT:\t%08x\n",
> >   			   I915_READ(DPINVGTT));
> > -		intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, pref);
> > +		intel_display_power_put(i915, POWER_DOMAIN_INIT, pref);
> >   		for (i = 0; i < 4; i++) {
> >   			seq_printf(m, "GT Interrupt IMR %d:\t%08x\n",
> > @@ -544,7 +544,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			   I915_READ(GEN8_PCU_IIR));
> >   		seq_printf(m, "PCU interrupt enable:\t%08x\n",
> >   			   I915_READ(GEN8_PCU_IER));
> > -	} else if (INTEL_GEN(dev_priv) >= 11) {
> > +	} else if (INTEL_GEN(i915) >= 11) {
> >   		seq_printf(m, "Master Interrupt Control:  %08x\n",
> >   			   I915_READ(GEN11_GFX_MSTR_IRQ));
> > @@ -565,7 +565,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			   I915_READ(GEN11_DISPLAY_INT_CTL));
> >   		gen8_display_interrupt_info(m);
> > -	} else if (INTEL_GEN(dev_priv) >= 8) {
> > +	} else if (INTEL_GEN(i915) >= 8) {
> >   		seq_printf(m, "Master Interrupt Control:\t%08x\n",
> >   			   I915_READ(GEN8_MASTER_IRQ));
> > @@ -579,7 +579,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   		}
> >   		gen8_display_interrupt_info(m);
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		seq_printf(m, "Display IER:\t%08x\n",
> >   			   I915_READ(VLV_IER));
> >   		seq_printf(m, "Display IIR:\t%08x\n",
> > @@ -588,12 +588,12 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			   I915_READ(VLV_IIR_RW));
> >   		seq_printf(m, "Display IMR:\t%08x\n",
> >   			   I915_READ(VLV_IMR));
> > -		for_each_pipe(dev_priv, pipe) {
> > +		for_each_pipe(i915, pipe) {
> >   			enum intel_display_power_domain power_domain;
> >   			intel_wakeref_t pref;
> >   			power_domain = POWER_DOMAIN_PIPE(pipe);
> > -			pref = intel_display_power_get_if_enabled(dev_priv,
> > +			pref = intel_display_power_get_if_enabled(i915,
> >   								  power_domain);
> >   			if (!pref) {
> >   				seq_printf(m, "Pipe %c power disabled\n",
> > @@ -604,7 +604,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			seq_printf(m, "Pipe %c stat:\t%08x\n",
> >   				   pipe_name(pipe),
> >   				   I915_READ(PIPESTAT(pipe)));
> > -			intel_display_power_put(dev_priv, power_domain, pref);
> > +			intel_display_power_put(i915, power_domain, pref);
> >   		}
> >   		seq_printf(m, "Master IER:\t%08x\n",
> > @@ -631,14 +631,14 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   		seq_printf(m, "DPINVGTT:\t%08x\n",
> >   			   I915_READ(DPINVGTT));
> > -	} else if (!HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (!HAS_PCH_SPLIT(i915)) {
> >   		seq_printf(m, "Interrupt enable:    %08x\n",
> >   			   I915_READ(GEN2_IER));
> >   		seq_printf(m, "Interrupt identity:  %08x\n",
> >   			   I915_READ(GEN2_IIR));
> >   		seq_printf(m, "Interrupt mask:      %08x\n",
> >   			   I915_READ(GEN2_IMR));
> > -		for_each_pipe(dev_priv, pipe)
> > +		for_each_pipe(i915, pipe)
> >   			seq_printf(m, "Pipe %c stat:         %08x\n",
> >   				   pipe_name(pipe),
> >   				   I915_READ(PIPESTAT(pipe)));
> > @@ -663,7 +663,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   			   I915_READ(GTIMR));
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		seq_printf(m, "RCS Intr Mask:\t %08x\n",
> >   			   I915_READ(GEN11_RCS0_RSVD_INTR_MASK));
> >   		seq_printf(m, "BCS Intr Mask:\t %08x\n",
> > @@ -683,35 +683,35 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
> >   		seq_printf(m, "Gunit/CSME Intr Mask:\t %08x\n",
> >   			   I915_READ(GEN11_GUNIT_CSME_INTR_MASK));
> > -	} else if (INTEL_GEN(dev_priv) >= 6) {
> > -		for_each_engine(engine, dev_priv, id) {
> > +	} else if (INTEL_GEN(i915) >= 6) {
> > +		for_each_engine(engine, i915, id) {
> >   			seq_printf(m,
> >   				   "Graphics Interrupt mask (%s):	%08x\n",
> >   				   engine->name, ENGINE_READ(engine, RING_IMR));
> >   		}
> >   	}
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_gem_fence_regs_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	int i, ret;
> >   	ret = mutex_lock_interruptible(&dev->struct_mutex);
> >   	if (ret)
> >   		return ret;
> > -	seq_printf(m, "Total fences = %d\n", dev_priv->num_fence_regs);
> > -	for (i = 0; i < dev_priv->num_fence_regs; i++) {
> > -		struct i915_vma *vma = dev_priv->fence_regs[i].vma;
> > +	seq_printf(m, "Total fences = %d\n", i915->num_fence_regs);
> > +	for (i = 0; i < i915->num_fence_regs; i++) {
> > +		struct i915_vma *vma = i915->fence_regs[i].vma;
> >   		seq_printf(m, "Fence %d, pin count = %d, object = ",
> > -			   i, dev_priv->fence_regs[i].pin_count);
> > +			   i, i915->fence_regs[i].pin_count);
> >   		if (!vma)
> >   			seq_puts(m, "unused");
> >   		else
> > @@ -825,15 +825,15 @@ static const struct file_operations i915_error_state_fops = {
> >   static int i915_frequency_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct intel_uncore *uncore = &i915->uncore;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	intel_wakeref_t wakeref;
> >   	int ret = 0;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	if (IS_GEN(dev_priv, 5)) {
> > +	if (IS_GEN(i915, 5)) {
> >   		u16 rgvswctl = intel_uncore_read16(uncore, MEMSWCTL);
> >   		u16 rgvstat = intel_uncore_read16(uncore, MEMSTAT_ILK);
> > @@ -843,7 +843,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   			   MEMSTAT_VID_SHIFT);
> >   		seq_printf(m, "Current P-state: %d\n",
> >   			   (rgvstat & MEMSTAT_PSTATE_MASK) >> MEMSTAT_PSTATE_SHIFT);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		u32 rpmodectl, freq_sts;
> >   		rpmodectl = I915_READ(GEN6_RP_CONTROL);
> > @@ -855,32 +855,32 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   			   yesno((rpmodectl & GEN6_RP_MEDIA_MODE_MASK) ==
> >   				  GEN6_RP_MEDIA_SW_MODE));
> > -		vlv_punit_get(dev_priv);
> > -		freq_sts = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> > -		vlv_punit_put(dev_priv);
> > +		vlv_punit_get(i915);
> > +		freq_sts = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> > +		vlv_punit_put(i915);
> >   		seq_printf(m, "PUNIT_REG_GPU_FREQ_STS: 0x%08x\n", freq_sts);
> > -		seq_printf(m, "DDR freq: %d MHz\n", dev_priv->mem_freq);
> > +		seq_printf(m, "DDR freq: %d MHz\n", i915->mem_freq);
> >   		seq_printf(m, "actual GPU freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, (freq_sts >> 8) & 0xff));
> > +			   intel_gpu_freq(i915, (freq_sts >> 8) & 0xff));
> >   		seq_printf(m, "current GPU freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->cur_freq));
> > +			   intel_gpu_freq(i915, rps->cur_freq));
> >   		seq_printf(m, "max GPU freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->max_freq));
> > +			   intel_gpu_freq(i915, rps->max_freq));
> >   		seq_printf(m, "min GPU freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->min_freq));
> > +			   intel_gpu_freq(i915, rps->min_freq));
> >   		seq_printf(m, "idle GPU freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->idle_freq));
> > +			   intel_gpu_freq(i915, rps->idle_freq));
> >   		seq_printf(m,
> >   			   "efficient (RPe) frequency: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->efficient_freq));
> > -	} else if (INTEL_GEN(dev_priv) >= 6) {
> > +			   intel_gpu_freq(i915, rps->efficient_freq));
> > +	} else if (INTEL_GEN(i915) >= 6) {
> >   		u32 rp_state_limits;
> >   		u32 gt_perf_status;
> >   		u32 rp_state_cap;
> > @@ -892,7 +892,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   		int max_freq;
> >   		rp_state_limits = I915_READ(GEN6_RP_STATE_LIMITS);
> > -		if (IS_GEN9_LP(dev_priv)) {
> > +		if (IS_GEN9_LP(i915)) {
> >   			rp_state_cap = I915_READ(BXT_RP_STATE_CAP);
> >   			gt_perf_status = I915_READ(BXT_GT_PERF_STATUS);
> >   		} else {
> > @@ -901,19 +901,19 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   		}
> >   		/* RPSTAT1 is in the GT power well */
> > -		intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +		intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   		reqf = I915_READ(GEN6_RPNSWREQ);
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			reqf >>= 23;
> >   		else {
> >   			reqf &= ~GEN6_TURBO_DISABLE;
> > -			if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +			if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   				reqf >>= 24;
> >   			else
> >   				reqf >>= 25;
> >   		}
> > -		reqf = intel_gpu_freq(dev_priv, reqf);
> > +		reqf = intel_gpu_freq(i915, reqf);
> >   		rpmodectl = I915_READ(GEN6_RP_CONTROL);
> >   		rpinclimit = I915_READ(GEN6_RP_UP_THRESHOLD);
> > @@ -926,12 +926,12 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   		rpdownei = I915_READ(GEN6_RP_CUR_DOWN_EI) & GEN6_CURIAVG_MASK;
> >   		rpcurdown = I915_READ(GEN6_RP_CUR_DOWN) & GEN6_CURBSYTAVG_MASK;
> >   		rpprevdown = I915_READ(GEN6_RP_PREV_DOWN) & GEN6_CURBSYTAVG_MASK;
> > -		cagf = intel_gpu_freq(dev_priv,
> > -				      intel_get_cagf(dev_priv, rpstat));
> > +		cagf = intel_gpu_freq(i915,
> > +				      intel_get_cagf(i915, rpstat));
> > -		intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +		intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > -		if (INTEL_GEN(dev_priv) >= 11) {
> > +		if (INTEL_GEN(i915) >= 11) {
> >   			pm_ier = I915_READ(GEN11_GPM_WGBOXPERF_INTR_ENABLE);
> >   			pm_imr = I915_READ(GEN11_GPM_WGBOXPERF_INTR_MASK);
> >   			/*
> > @@ -940,7 +940,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   			 */
> >   			pm_isr = 0;
> >   			pm_iir = 0;
> > -		} else if (INTEL_GEN(dev_priv) >= 8) {
> > +		} else if (INTEL_GEN(i915) >= 8) {
> >   			pm_ier = I915_READ(GEN8_GT_IER(2));
> >   			pm_imr = I915_READ(GEN8_GT_IMR(2));
> >   			pm_isr = I915_READ(GEN8_GT_ISR(2));
> > @@ -963,14 +963,14 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   		seq_printf(m, "PM IER=0x%08x IMR=0x%08x, MASK=0x%08x\n",
> >   			   pm_ier, pm_imr, pm_mask);
> > -		if (INTEL_GEN(dev_priv) <= 10)
> > +		if (INTEL_GEN(i915) <= 10)
> >   			seq_printf(m, "PM ISR=0x%08x IIR=0x%08x\n",
> >   				   pm_isr, pm_iir);
> >   		seq_printf(m, "pm_intrmsk_mbz: 0x%08x\n",
> >   			   rps->pm_intrmsk_mbz);
> >   		seq_printf(m, "GT_PERF_STATUS: 0x%08x\n", gt_perf_status);
> >   		seq_printf(m, "Render p-state ratio: %d\n",
> > -			   (gt_perf_status & (INTEL_GEN(dev_priv) >= 9 ? 0x1ff00 : 0xff00)) >> 8);
> > +			   (gt_perf_status & (INTEL_GEN(i915) >= 9 ? 0x1ff00 : 0xff00)) >> 8);
> >   		seq_printf(m, "Render p-state VID: %d\n",
> >   			   gt_perf_status & 0xff);
> >   		seq_printf(m, "Render p-state limit: %d\n",
> > @@ -982,72 +982,72 @@ static int i915_frequency_info(struct seq_file *m, void *unused)
> >   		seq_printf(m, "RPNSWREQ: %dMHz\n", reqf);
> >   		seq_printf(m, "CAGF: %dMHz\n", cagf);
> >   		seq_printf(m, "RP CUR UP EI: %d (%dus)\n",
> > -			   rpupei, GT_PM_INTERVAL_TO_US(dev_priv, rpupei));
> > +			   rpupei, GT_PM_INTERVAL_TO_US(i915, rpupei));
> >   		seq_printf(m, "RP CUR UP: %d (%dus)\n",
> > -			   rpcurup, GT_PM_INTERVAL_TO_US(dev_priv, rpcurup));
> > +			   rpcurup, GT_PM_INTERVAL_TO_US(i915, rpcurup));
> >   		seq_printf(m, "RP PREV UP: %d (%dus)\n",
> > -			   rpprevup, GT_PM_INTERVAL_TO_US(dev_priv, rpprevup));
> > +			   rpprevup, GT_PM_INTERVAL_TO_US(i915, rpprevup));
> >   		seq_printf(m, "Up threshold: %d%%\n",
> >   			   rps->power.up_threshold);
> >   		seq_printf(m, "RP CUR DOWN EI: %d (%dus)\n",
> > -			   rpdownei, GT_PM_INTERVAL_TO_US(dev_priv, rpdownei));
> > +			   rpdownei, GT_PM_INTERVAL_TO_US(i915, rpdownei));
> >   		seq_printf(m, "RP CUR DOWN: %d (%dus)\n",
> > -			   rpcurdown, GT_PM_INTERVAL_TO_US(dev_priv, rpcurdown));
> > +			   rpcurdown, GT_PM_INTERVAL_TO_US(i915, rpcurdown));
> >   		seq_printf(m, "RP PREV DOWN: %d (%dus)\n",
> > -			   rpprevdown, GT_PM_INTERVAL_TO_US(dev_priv, rpprevdown));
> > +			   rpprevdown, GT_PM_INTERVAL_TO_US(i915, rpprevdown));
> >   		seq_printf(m, "Down threshold: %d%%\n",
> >   			   rps->power.down_threshold);
> > -		max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 0 :
> > +		max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 0 :
> >   			    rp_state_cap >> 16) & 0xff;
> > -		max_freq *= (IS_GEN9_BC(dev_priv) ||
> > -			     INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> > +		max_freq *= (IS_GEN9_BC(i915) ||
> > +			     INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> >   		seq_printf(m, "Lowest (RPN) frequency: %dMHz\n",
> > -			   intel_gpu_freq(dev_priv, max_freq));
> > +			   intel_gpu_freq(i915, max_freq));
> >   		max_freq = (rp_state_cap & 0xff00) >> 8;
> > -		max_freq *= (IS_GEN9_BC(dev_priv) ||
> > -			     INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> > +		max_freq *= (IS_GEN9_BC(i915) ||
> > +			     INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> >   		seq_printf(m, "Nominal (RP1) frequency: %dMHz\n",
> > -			   intel_gpu_freq(dev_priv, max_freq));
> > +			   intel_gpu_freq(i915, max_freq));
> > -		max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 16 :
> > +		max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 16 :
> >   			    rp_state_cap >> 0) & 0xff;
> > -		max_freq *= (IS_GEN9_BC(dev_priv) ||
> > -			     INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1);
> > +		max_freq *= (IS_GEN9_BC(i915) ||
> > +			     INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1);
> >   		seq_printf(m, "Max non-overclocked (RP0) frequency: %dMHz\n",
> > -			   intel_gpu_freq(dev_priv, max_freq));
> > +			   intel_gpu_freq(i915, max_freq));
> >   		seq_printf(m, "Max overclocked frequency: %dMHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->max_freq));
> > +			   intel_gpu_freq(i915, rps->max_freq));
> >   		seq_printf(m, "Current freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->cur_freq));
> > +			   intel_gpu_freq(i915, rps->cur_freq));
> >   		seq_printf(m, "Actual freq: %d MHz\n", cagf);
> >   		seq_printf(m, "Idle freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->idle_freq));
> > +			   intel_gpu_freq(i915, rps->idle_freq));
> >   		seq_printf(m, "Min freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->min_freq));
> > +			   intel_gpu_freq(i915, rps->min_freq));
> >   		seq_printf(m, "Boost freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->boost_freq));
> > +			   intel_gpu_freq(i915, rps->boost_freq));
> >   		seq_printf(m, "Max freq: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->max_freq));
> > +			   intel_gpu_freq(i915, rps->max_freq));
> >   		seq_printf(m,
> >   			   "efficient (RPe) frequency: %d MHz\n",
> > -			   intel_gpu_freq(dev_priv, rps->efficient_freq));
> > +			   intel_gpu_freq(i915, rps->efficient_freq));
> >   	} else {
> >   		seq_puts(m, "no P-state info available\n");
> >   	}
> > -	seq_printf(m, "Current CD clock frequency: %d kHz\n", dev_priv->cdclk.hw.cdclk);
> > -	seq_printf(m, "Max CD clock frequency: %d kHz\n", dev_priv->max_cdclk_freq);
> > -	seq_printf(m, "Max pixel clock frequency: %d kHz\n", dev_priv->max_dotclk_freq);
> > +	seq_printf(m, "Current CD clock frequency: %d kHz\n", i915->cdclk.hw.cdclk);
> > +	seq_printf(m, "Max CD clock frequency: %d kHz\n", i915->max_cdclk_freq);
> > +	seq_printf(m, "Max pixel clock frequency: %d kHz\n", i915->max_dotclk_freq);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return ret;
> >   }
> > -static void i915_instdone_info(struct drm_i915_private *dev_priv,
> > +static void i915_instdone_info(struct drm_i915_private *i915,
> >   			       struct seq_file *m,
> >   			       struct intel_instdone *instdone)
> >   {
> > @@ -1057,37 +1057,37 @@ static void i915_instdone_info(struct drm_i915_private *dev_priv,
> >   	seq_printf(m, "\t\tINSTDONE: 0x%08x\n",
> >   		   instdone->instdone);
> > -	if (INTEL_GEN(dev_priv) <= 3)
> > +	if (INTEL_GEN(i915) <= 3)
> >   		return;
> >   	seq_printf(m, "\t\tSC_INSTDONE: 0x%08x\n",
> >   		   instdone->slice_common);
> > -	if (INTEL_GEN(dev_priv) <= 6)
> > +	if (INTEL_GEN(i915) <= 6)
> >   		return;
> > -	for_each_instdone_slice_subslice(dev_priv, slice, subslice)
> > +	for_each_instdone_slice_subslice(i915, slice, subslice)
> >   		seq_printf(m, "\t\tSAMPLER_INSTDONE[%d][%d]: 0x%08x\n",
> >   			   slice, subslice, instdone->sampler[slice][subslice]);
> > -	for_each_instdone_slice_subslice(dev_priv, slice, subslice)
> > +	for_each_instdone_slice_subslice(i915, slice, subslice)
> >   		seq_printf(m, "\t\tROW_INSTDONE[%d][%d]: 0x%08x\n",
> >   			   slice, subslice, instdone->row[slice][subslice]);
> >   }
> >   static int i915_hangcheck_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	struct intel_engine_cs *engine;
> >   	u64 acthd[I915_NUM_ENGINES];
> >   	struct intel_instdone instdone;
> >   	intel_wakeref_t wakeref;
> >   	enum intel_engine_id id;
> > -	seq_printf(m, "Reset flags: %lx\n", dev_priv->gpu_error.flags);
> > -	if (test_bit(I915_WEDGED, &dev_priv->gpu_error.flags))
> > +	seq_printf(m, "Reset flags: %lx\n", i915->gpu_error.flags);
> > +	if (test_bit(I915_WEDGED, &i915->gpu_error.flags))
> >   		seq_puts(m, "\tWedged\n");
> > -	if (test_bit(I915_RESET_BACKOFF, &dev_priv->gpu_error.flags))
> > +	if (test_bit(I915_RESET_BACKOFF, &i915->gpu_error.flags))
> >   		seq_puts(m, "\tDevice (global) reset in progress\n");
> >   	if (!i915_modparams.enable_hangcheck) {
> > @@ -1095,25 +1095,25 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
> >   		return 0;
> >   	}
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > -		for_each_engine(engine, dev_priv, id)
> > +	with_intel_runtime_pm(i915, wakeref) {
> > +		for_each_engine(engine, i915, id)
> >   			acthd[id] = intel_engine_get_active_head(engine);
> > -		intel_engine_get_instdone(dev_priv->engine[RCS0], &instdone);
> > +		intel_engine_get_instdone(i915->engine[RCS0], &instdone);
> >   	}
> > -	if (timer_pending(&dev_priv->gpu_error.hangcheck_work.timer))
> > +	if (timer_pending(&i915->gpu_error.hangcheck_work.timer))
> >   		seq_printf(m, "Hangcheck active, timer fires in %dms\n",
> > -			   jiffies_to_msecs(dev_priv->gpu_error.hangcheck_work.timer.expires -
> > +			   jiffies_to_msecs(i915->gpu_error.hangcheck_work.timer.expires -
> >   					    jiffies));
> > -	else if (delayed_work_pending(&dev_priv->gpu_error.hangcheck_work))
> > +	else if (delayed_work_pending(&i915->gpu_error.hangcheck_work))
> >   		seq_puts(m, "Hangcheck active, work pending\n");
> >   	else
> >   		seq_puts(m, "Hangcheck inactive\n");
> > -	seq_printf(m, "GT active? %s\n", yesno(dev_priv->gt.awake));
> > +	seq_printf(m, "GT active? %s\n", yesno(i915->gt.awake));
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		seq_printf(m, "%s: %d ms ago\n",
> >   			   engine->name,
> >   			   jiffies_to_msecs(jiffies -
> > @@ -1126,11 +1126,11 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
> >   		if (engine->id == RCS0) {
> >   			seq_puts(m, "\tinstdone read =\n");
> > -			i915_instdone_info(dev_priv, m, &instdone);
> > +			i915_instdone_info(i915, m, &instdone);
> >   			seq_puts(m, "\tinstdone accu =\n");
> > -			i915_instdone_info(dev_priv, m,
> > +			i915_instdone_info(i915, m,
> >   					   &engine->hangcheck.instdone);
> >   		}
> >   	}
> > @@ -1140,14 +1140,14 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused)
> >   static int i915_reset_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct i915_gpu_error *error = &dev_priv->gpu_error;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct i915_gpu_error *error = &i915->gpu_error;
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	seq_printf(m, "full gpu reset = %u\n", i915_reset_count(error));
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		seq_printf(m, "%s = %u\n", engine->name,
> >   			   i915_reset_engine_count(error, engine));
> >   	}
> > @@ -1235,16 +1235,16 @@ static void print_rc6_res(struct seq_file *m,
> >   			  const char *title,
> >   			  const i915_reg_t reg)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	seq_printf(m, "%s %u (%llu us)\n",
> >   		   title, I915_READ(reg),
> > -		   intel_rc6_residency_us(dev_priv, reg));
> > +		   intel_rc6_residency_us(i915, reg));
> >   }
> >   static int vlv_drpc_info(struct seq_file *m)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	u32 rcctl1, pw_status;
> >   	pw_status = I915_READ(VLV_GTLC_PW_STATUS);
> > @@ -1266,7 +1266,7 @@ static int vlv_drpc_info(struct seq_file *m)
> >   static int gen6_drpc_info(struct seq_file *m)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	u32 gt_core_status, rcctl1, rc6vids = 0;
> >   	u32 gen9_powergate_enable = 0, gen9_powergate_status = 0;
> > @@ -1274,20 +1274,20 @@ static int gen6_drpc_info(struct seq_file *m)
> >   	trace_i915_reg_rw(false, GEN6_GT_CORE_STATUS, gt_core_status, 4, true);
> >   	rcctl1 = I915_READ(GEN6_RC_CONTROL);
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		gen9_powergate_enable = I915_READ(GEN9_PG_ENABLE);
> >   		gen9_powergate_status = I915_READ(GEN9_PWRGT_DOMAIN_STATUS);
> >   	}
> > -	if (INTEL_GEN(dev_priv) <= 7)
> > -		sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS,
> > +	if (INTEL_GEN(i915) <= 7)
> > +		sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS,
> >   				       &rc6vids, NULL);
> >   	seq_printf(m, "RC1e Enabled: %s\n",
> >   		   yesno(rcctl1 & GEN6_RC_CTL_RC1e_ENABLE));
> >   	seq_printf(m, "RC6 Enabled: %s\n",
> >   		   yesno(rcctl1 & GEN6_RC_CTL_RC6_ENABLE));
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		seq_printf(m, "Render Well Gating Enabled: %s\n",
> >   			yesno(gen9_powergate_enable & GEN9_RENDER_PG_ENABLE));
> >   		seq_printf(m, "Media Well Gating Enabled: %s\n",
> > @@ -1321,7 +1321,7 @@ static int gen6_drpc_info(struct seq_file *m)
> >   	seq_printf(m, "Core Power Down: %s\n",
> >   		   yesno(gt_core_status & GEN6_CORE_CPD_STATE_MASK));
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		seq_printf(m, "Render Power Well: %s\n",
> >   			(gen9_powergate_status &
> >   			 GEN9_PWRGT_RENDER_STATUS_MASK) ? "Up" : "Down");
> > @@ -1337,7 +1337,7 @@ static int gen6_drpc_info(struct seq_file *m)
> >   	print_rc6_res(m, "RC6+ residency since boot:", GEN6_GT_GFX_RC6p);
> >   	print_rc6_res(m, "RC6++ residency since boot:", GEN6_GT_GFX_RC6pp);
> > -	if (INTEL_GEN(dev_priv) <= 7) {
> > +	if (INTEL_GEN(i915) <= 7) {
> >   		seq_printf(m, "RC6   voltage: %dmV\n",
> >   			   GEN6_DECODE_RC6_VID(((rc6vids >> 0) & 0xff)));
> >   		seq_printf(m, "RC6+  voltage: %dmV\n",
> > @@ -1351,14 +1351,14 @@ static int gen6_drpc_info(struct seq_file *m)
> >   static int i915_drpc_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> >   	int err = -ENODEV;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	with_intel_runtime_pm(i915, wakeref) {
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   			err = vlv_drpc_info(m);
> > -		else if (INTEL_GEN(dev_priv) >= 6)
> > +		else if (INTEL_GEN(i915) >= 6)
> >   			err = gen6_drpc_info(m);
> >   		else
> >   			err = ironlake_drpc_info(m);
> > @@ -1369,44 +1369,44 @@ static int i915_drpc_info(struct seq_file *m, void *unused)
> >   static int i915_frontbuffer_tracking(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	seq_printf(m, "FB tracking busy bits: 0x%08x\n",
> > -		   dev_priv->fb_tracking.busy_bits);
> > +		   i915->fb_tracking.busy_bits);
> >   	seq_printf(m, "FB tracking flip bits: 0x%08x\n",
> > -		   dev_priv->fb_tracking.flip_bits);
> > +		   i915->fb_tracking.flip_bits);
> >   	return 0;
> >   }
> >   static int i915_fbc_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	intel_wakeref_t wakeref;
> > -	if (!HAS_FBC(dev_priv))
> > +	if (!HAS_FBC(i915))
> >   		return -ENODEV;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	mutex_lock(&fbc->lock);
> > -	if (intel_fbc_is_active(dev_priv))
> > +	if (intel_fbc_is_active(i915))
> >   		seq_puts(m, "FBC enabled\n");
> >   	else
> >   		seq_printf(m, "FBC disabled: %s\n", fbc->no_fbc_reason);
> > -	if (intel_fbc_is_active(dev_priv)) {
> > +	if (intel_fbc_is_active(i915)) {
> >   		u32 mask;
> > -		if (INTEL_GEN(dev_priv) >= 8)
> > +		if (INTEL_GEN(i915) >= 8)
> >   			mask = I915_READ(IVB_FBC_STATUS2) & BDW_FBC_COMP_SEG_MASK;
> > -		else if (INTEL_GEN(dev_priv) >= 7)
> > +		else if (INTEL_GEN(i915) >= 7)
> >   			mask = I915_READ(IVB_FBC_STATUS2) & IVB_FBC_COMP_SEG_MASK;
> > -		else if (INTEL_GEN(dev_priv) >= 5)
> > +		else if (INTEL_GEN(i915) >= 5)
> >   			mask = I915_READ(ILK_DPFC_STATUS) & ILK_DPFC_COMP_SEG_MASK;
> > -		else if (IS_G4X(dev_priv))
> > +		else if (IS_G4X(i915))
> >   			mask = I915_READ(DPFC_STATUS) & DPFC_COMP_SEG_MASK;
> >   		else
> >   			mask = I915_READ(FBC_STATUS) & (FBC_STAT_COMPRESSING |
> > @@ -1416,41 +1416,41 @@ static int i915_fbc_status(struct seq_file *m, void *unused)
> >   	}
> >   	mutex_unlock(&fbc->lock);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_fbc_false_color_get(void *data, u64 *val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> > -	if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv))
> > +	if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915))
> >   		return -ENODEV;
> > -	*val = dev_priv->fbc.false_color;
> > +	*val = i915->fbc.false_color;
> >   	return 0;
> >   }
> >   static int i915_fbc_false_color_set(void *data, u64 val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> >   	u32 reg;
> > -	if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv))
> > +	if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915))
> >   		return -ENODEV;
> > -	mutex_lock(&dev_priv->fbc.lock);
> > +	mutex_lock(&i915->fbc.lock);
> >   	reg = I915_READ(ILK_DPFC_CONTROL);
> > -	dev_priv->fbc.false_color = val;
> > +	i915->fbc.false_color = val;
> >   	I915_WRITE(ILK_DPFC_CONTROL, val ?
> >   		   (reg | FBC_CTL_FALSE_COLOR) :
> >   		   (reg & ~FBC_CTL_FALSE_COLOR));
> > -	mutex_unlock(&dev_priv->fbc.lock);
> > +	mutex_unlock(&i915->fbc.lock);
> >   	return 0;
> >   }
> > @@ -1460,18 +1460,18 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_fbc_false_color_fops,
> >   static int i915_ips_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> > -	if (!HAS_IPS(dev_priv))
> > +	if (!HAS_IPS(i915))
> >   		return -ENODEV;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	seq_printf(m, "Enabled by kernel parameter: %s\n",
> >   		   yesno(i915_modparams.enable_ips));
> > -	if (INTEL_GEN(dev_priv) >= 8) {
> > +	if (INTEL_GEN(i915) >= 8) {
> >   		seq_puts(m, "Currently: unknown\n");
> >   	} else {
> >   		if (I915_READ(IPS_CTL) & IPS_ENABLE)
> > @@ -1480,34 +1480,34 @@ static int i915_ips_status(struct seq_file *m, void *unused)
> >   			seq_puts(m, "Currently: disabled\n");
> >   	}
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_sr_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> >   	bool sr_enabled = false;
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		/* no global SR status; inspect per-plane WM */;
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > +	else if (HAS_PCH_SPLIT(i915))
> >   		sr_enabled = I915_READ(WM1_LP_ILK) & WM1_LP_SR_EN;
> > -	else if (IS_I965GM(dev_priv) || IS_G4X(dev_priv) ||
> > -		 IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> > +	else if (IS_I965GM(i915) || IS_G4X(i915) ||
> > +		 IS_I945G(i915) || IS_I945GM(i915))
> >   		sr_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> > -	else if (IS_I915GM(dev_priv))
> > +	else if (IS_I915GM(i915))
> >   		sr_enabled = I915_READ(INSTPM) & INSTPM_SELF_EN;
> > -	else if (IS_PINEVIEW(dev_priv))
> > +	else if (IS_PINEVIEW(i915))
> >   		sr_enabled = I915_READ(DSPFW3) & PINEVIEW_SELF_REFRESH_EN;
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		sr_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
> >   	seq_printf(m, "self-refresh: %s\n", enableddisabled(sr_enabled));
> > @@ -1540,18 +1540,18 @@ static int i915_emon_status(struct seq_file *m, void *unused)
> >   static int i915_ring_freq_table(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	unsigned int max_gpu_freq, min_gpu_freq;
> >   	intel_wakeref_t wakeref;
> >   	int gpu_freq, ia_freq;
> > -	if (!HAS_LLC(dev_priv))
> > +	if (!HAS_LLC(i915))
> >   		return -ENODEV;
> >   	min_gpu_freq = rps->min_freq;
> >   	max_gpu_freq = rps->max_freq;
> > -	if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +	if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> >   		/* Convert GT frequency to 50 HZ units */
> >   		min_gpu_freq /= GEN9_FREQ_SCALER;
> >   		max_gpu_freq /= GEN9_FREQ_SCALER;
> > @@ -1559,30 +1559,30 @@ static int i915_ring_freq_table(struct seq_file *m, void *unused)
> >   	seq_puts(m, "GPU freq (MHz)\tEffective CPU freq (MHz)\tEffective Ring freq (MHz)\n");
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	for (gpu_freq = min_gpu_freq; gpu_freq <= max_gpu_freq; gpu_freq++) {
> >   		ia_freq = gpu_freq;
> > -		sandybridge_pcode_read(dev_priv,
> > +		sandybridge_pcode_read(i915,
> >   				       GEN6_PCODE_READ_MIN_FREQ_TABLE,
> >   				       &ia_freq, NULL);
> >   		seq_printf(m, "%d\t\t%d\t\t\t\t%d\n",
> > -			   intel_gpu_freq(dev_priv, (gpu_freq *
> > -						     (IS_GEN9_BC(dev_priv) ||
> > -						      INTEL_GEN(dev_priv) >= 10 ?
> > +			   intel_gpu_freq(i915, (gpu_freq *
> > +						     (IS_GEN9_BC(i915) ||
> > +						      INTEL_GEN(i915) >= 10 ?
> >   						      GEN9_FREQ_SCALER : 1))),
> >   			   ((ia_freq >> 0) & 0xff) * 100,
> >   			   ((ia_freq >> 8) & 0xff) * 100);
> >   	}
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_opregion(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> > +	struct intel_opregion *opregion = &i915->opregion;
> >   	int ret;
> >   	ret = mutex_lock_interruptible(&dev->struct_mutex);
> > @@ -1610,8 +1610,8 @@ static int i915_vbt(struct seq_file *m, void *unused)
> >   static int i915_gem_framebuffer_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_framebuffer *fbdev_fb = NULL;
> >   	struct drm_framebuffer *drm_fb;
> >   	int ret;
> > @@ -1621,8 +1621,8 @@ static int i915_gem_framebuffer_info(struct seq_file *m, void *data)
> >   		return ret;
> >   #ifdef CONFIG_DRM_FBDEV_EMULATION
> > -	if (dev_priv->fbdev && dev_priv->fbdev->helper.fb) {
> > -		fbdev_fb = to_intel_framebuffer(dev_priv->fbdev->helper.fb);
> > +	if (i915->fbdev && i915->fbdev->helper.fb) {
> > +		fbdev_fb = to_intel_framebuffer(i915->fbdev->helper.fb);
> >   		seq_printf(m, "fbcon size: %d x %d, depth %d, %d bpp, modifier 0x%llx, refcount %d, obj ",
> >   			   fbdev_fb->base.width,
> > @@ -1666,8 +1666,8 @@ static void describe_ctx_ring(struct seq_file *m, struct intel_ring *ring)
> >   static int i915_context_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct i915_gem_context *ctx;
> >   	int ret;
> > @@ -1675,7 +1675,7 @@ static int i915_context_status(struct seq_file *m, void *unused)
> >   	if (ret)
> >   		return ret;
> > -	list_for_each_entry(ctx, &dev_priv->contexts.list, link) {
> > +	list_for_each_entry(ctx, &i915->contexts.list, link) {
> >   		struct i915_gem_engines_iter it;
> >   		struct intel_context *ce;
> > @@ -1746,18 +1746,18 @@ static const char *swizzle_string(unsigned swizzle)
> >   static int i915_swizzle_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	intel_wakeref_t wakeref;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	seq_printf(m, "bit6 swizzle for X-tiling = %s\n",
> > -		   swizzle_string(dev_priv->mm.bit_6_swizzle_x));
> > +		   swizzle_string(i915->mm.bit_6_swizzle_x));
> >   	seq_printf(m, "bit6 swizzle for Y-tiling = %s\n",
> > -		   swizzle_string(dev_priv->mm.bit_6_swizzle_y));
> > +		   swizzle_string(i915->mm.bit_6_swizzle_y));
> > -	if (IS_GEN_RANGE(dev_priv, 3, 4)) {
> > +	if (IS_GEN_RANGE(i915, 3, 4)) {
> >   		seq_printf(m, "DDC = 0x%08x\n",
> >   			   intel_uncore_read(uncore, DCC));
> >   		seq_printf(m, "DDC2 = 0x%08x\n",
> > @@ -1766,7 +1766,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> >   			   intel_uncore_read16(uncore, C0DRB3));
> >   		seq_printf(m, "C1DRB3 = 0x%04x\n",
> >   			   intel_uncore_read16(uncore, C1DRB3));
> > -	} else if (INTEL_GEN(dev_priv) >= 6) {
> > +	} else if (INTEL_GEN(i915) >= 6) {
> >   		seq_printf(m, "MAD_DIMM_C0 = 0x%08x\n",
> >   			   intel_uncore_read(uncore, MAD_DIMM_C0));
> >   		seq_printf(m, "MAD_DIMM_C1 = 0x%08x\n",
> > @@ -1775,7 +1775,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> >   			   intel_uncore_read(uncore, MAD_DIMM_C2));
> >   		seq_printf(m, "TILECTL = 0x%08x\n",
> >   			   intel_uncore_read(uncore, TILECTL));
> > -		if (INTEL_GEN(dev_priv) >= 8)
> > +		if (INTEL_GEN(i915) >= 8)
> >   			seq_printf(m, "GAMTARBMODE = 0x%08x\n",
> >   				   intel_uncore_read(uncore, GAMTARBMODE));
> >   		else
> > @@ -1785,10 +1785,10 @@ static int i915_swizzle_info(struct seq_file *m, void *data)
> >   			   intel_uncore_read(uncore, DISP_ARB_CTL));
> >   	}
> > -	if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> > +	if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES)
> >   		seq_puts(m, "L-shaped memory detected\n");
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> > @@ -1809,54 +1809,54 @@ static const char *rps_power_to_str(unsigned int power)
> >   static int i915_rps_boost_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 act_freq = rps->cur_freq;
> >   	intel_wakeref_t wakeref;
> > -	with_intel_runtime_pm_if_in_use(dev_priv, wakeref) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -			vlv_punit_get(dev_priv);
> > -			act_freq = vlv_punit_read(dev_priv,
> > +	with_intel_runtime_pm_if_in_use(i915, wakeref) {
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +			vlv_punit_get(i915);
> > +			act_freq = vlv_punit_read(i915,
> >   						  PUNIT_REG_GPU_FREQ_STS);
> > -			vlv_punit_put(dev_priv);
> > +			vlv_punit_put(i915);
> >   			act_freq = (act_freq >> 8) & 0xff;
> >   		} else {
> > -			act_freq = intel_get_cagf(dev_priv,
> > +			act_freq = intel_get_cagf(i915,
> >   						  I915_READ(GEN6_RPSTAT1));
> >   		}
> >   	}
> >   	seq_printf(m, "RPS enabled? %d\n", rps->enabled);
> > -	seq_printf(m, "GPU busy? %s\n", yesno(dev_priv->gt.awake));
> > +	seq_printf(m, "GPU busy? %s\n", yesno(i915->gt.awake));
> >   	seq_printf(m, "Boosts outstanding? %d\n",
> >   		   atomic_read(&rps->num_waiters));
> >   	seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
> >   	seq_printf(m, "Frequency requested %d, actual %d\n",
> > -		   intel_gpu_freq(dev_priv, rps->cur_freq),
> > -		   intel_gpu_freq(dev_priv, act_freq));
> > +		   intel_gpu_freq(i915, rps->cur_freq),
> > +		   intel_gpu_freq(i915, act_freq));
> >   	seq_printf(m, "  min hard:%d, soft:%d; max soft:%d, hard:%d\n",
> > -		   intel_gpu_freq(dev_priv, rps->min_freq),
> > -		   intel_gpu_freq(dev_priv, rps->min_freq_softlimit),
> > -		   intel_gpu_freq(dev_priv, rps->max_freq_softlimit),
> > -		   intel_gpu_freq(dev_priv, rps->max_freq));
> > +		   intel_gpu_freq(i915, rps->min_freq),
> > +		   intel_gpu_freq(i915, rps->min_freq_softlimit),
> > +		   intel_gpu_freq(i915, rps->max_freq_softlimit),
> > +		   intel_gpu_freq(i915, rps->max_freq));
> >   	seq_printf(m, "  idle:%d, efficient:%d, boost:%d\n",
> > -		   intel_gpu_freq(dev_priv, rps->idle_freq),
> > -		   intel_gpu_freq(dev_priv, rps->efficient_freq),
> > -		   intel_gpu_freq(dev_priv, rps->boost_freq));
> > +		   intel_gpu_freq(i915, rps->idle_freq),
> > +		   intel_gpu_freq(i915, rps->efficient_freq),
> > +		   intel_gpu_freq(i915, rps->boost_freq));
> >   	seq_printf(m, "Wait boosts: %d\n", atomic_read(&rps->boosts));
> > -	if (INTEL_GEN(dev_priv) >= 6 && rps->enabled && dev_priv->gt.awake) {
> > +	if (INTEL_GEN(i915) >= 6 && rps->enabled && i915->gt.awake) {
> >   		u32 rpup, rpupei;
> >   		u32 rpdown, rpdownei;
> > -		intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +		intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   		rpup = I915_READ_FW(GEN6_RP_CUR_UP) & GEN6_RP_EI_MASK;
> >   		rpupei = I915_READ_FW(GEN6_RP_CUR_UP_EI) & GEN6_RP_EI_MASK;
> >   		rpdown = I915_READ_FW(GEN6_RP_CUR_DOWN) & GEN6_RP_EI_MASK;
> >   		rpdownei = I915_READ_FW(GEN6_RP_CUR_DOWN_EI) & GEN6_RP_EI_MASK;
> > -		intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +		intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   		seq_printf(m, "\nRPS Autotuning (current \"%s\" window):\n",
> >   			   rps_power_to_str(rps->power.mode));
> > @@ -1875,29 +1875,29 @@ static int i915_rps_boost_info(struct seq_file *m, void *data)
> >   static int i915_llc(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	const bool edram = INTEL_GEN(dev_priv) > 8;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	const bool edram = INTEL_GEN(i915) > 8;
> > -	seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(dev_priv)));
> > +	seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(i915)));
> >   	seq_printf(m, "%s: %uMB\n", edram ? "eDRAM" : "eLLC",
> > -		   dev_priv->edram_size_mb);
> > +		   i915->edram_size_mb);
> >   	return 0;
> >   }
> >   static int i915_huc_load_status_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> >   	struct drm_printer p;
> > -	if (!HAS_HUC(dev_priv))
> > +	if (!HAS_HUC(i915))
> >   		return -ENODEV;
> >   	p = drm_seq_file_printer(m);
> > -	intel_uc_fw_dump(&dev_priv->huc.fw, &p);
> > +	intel_uc_fw_dump(&i915->huc.fw, &p);
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		seq_printf(m, "\nHuC status 0x%08x:\n", I915_READ(HUC_STATUS2));
> >   	return 0;
> > @@ -1905,17 +1905,17 @@ static int i915_huc_load_status_info(struct seq_file *m, void *data)
> >   static int i915_guc_load_status_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> >   	struct drm_printer p;
> > -	if (!HAS_GUC(dev_priv))
> > +	if (!HAS_GUC(i915))
> >   		return -ENODEV;
> >   	p = drm_seq_file_printer(m);
> > -	intel_uc_fw_dump(&dev_priv->guc.fw, &p);
> > +	intel_uc_fw_dump(&i915->guc.fw, &p);
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		u32 tmp = I915_READ(GUC_STATUS);
> >   		u32 i;
> > @@ -1954,9 +1954,9 @@ stringify_guc_log_type(enum guc_log_buffer_type type)
> >   }
> >   static void i915_guc_log_info(struct seq_file *m,
> > -			      struct drm_i915_private *dev_priv)
> > +			      struct drm_i915_private *i915)
> >   {
> > -	struct intel_guc_log *log = &dev_priv->guc.log;
> > +	struct intel_guc_log *log = &i915->guc.log;
> >   	enum guc_log_buffer_type type;
> >   	if (!intel_guc_log_relay_enabled(log)) {
> > @@ -1978,7 +1978,7 @@ static void i915_guc_log_info(struct seq_file *m,
> >   }
> >   static void i915_guc_client_info(struct seq_file *m,
> > -				 struct drm_i915_private *dev_priv,
> > +				 struct drm_i915_private *i915,
> >   				 struct intel_guc_client *client)
> >   {
> >   	struct intel_engine_cs *engine;
> > @@ -1990,7 +1990,7 @@ static void i915_guc_client_info(struct seq_file *m,
> >   	seq_printf(m, "\tDoorbell id %d, offset: 0x%lx\n",
> >   		client->doorbell_id, client->doorbell_offset);
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		u64 submissions = client->submissions[id];
> >   		tot += submissions;
> >   		seq_printf(m, "\tSubmissions: %llu %s\n",
> > @@ -2001,15 +2001,15 @@ static void i915_guc_client_info(struct seq_file *m,
> >   static int i915_guc_info(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	const struct intel_guc *guc = &dev_priv->guc;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	const struct intel_guc *guc = &i915->guc;
> > -	if (!USES_GUC(dev_priv))
> > +	if (!USES_GUC(i915))
> >   		return -ENODEV;
> > -	i915_guc_log_info(m, dev_priv);
> > +	i915_guc_log_info(m, i915);
> > -	if (!USES_GUC_SUBMISSION(dev_priv))
> > +	if (!USES_GUC_SUBMISSION(i915))
> >   		return 0;
> >   	GEM_BUG_ON(!guc->execbuf_client);
> > @@ -2019,11 +2019,11 @@ static int i915_guc_info(struct seq_file *m, void *data)
> >   	seq_printf(m, "Doorbell next cacheline: 0x%x\n", guc->db_cacheline);
> >   	seq_printf(m, "\nGuC execbuf client @ %p:\n", guc->execbuf_client);
> > -	i915_guc_client_info(m, dev_priv, guc->execbuf_client);
> > +	i915_guc_client_info(m, i915, guc->execbuf_client);
> >   	if (guc->preempt_client) {
> >   		seq_printf(m, "\nGuC preempt client @ %p:\n",
> >   			   guc->preempt_client);
> > -		i915_guc_client_info(m, dev_priv, guc->preempt_client);
> > +		i915_guc_client_info(m, i915, guc->preempt_client);
> >   	}
> >   	/* Add more as required ... */
> > @@ -2033,14 +2033,14 @@ static int i915_guc_info(struct seq_file *m, void *data)
> >   static int i915_guc_stage_pool(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	const struct intel_guc *guc = &dev_priv->guc;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	const struct intel_guc *guc = &i915->guc;
> >   	struct guc_stage_desc *desc = guc->stage_desc_pool_vaddr;
> >   	struct intel_guc_client *client = guc->execbuf_client;
> >   	intel_engine_mask_t tmp;
> >   	int index;
> > -	if (!USES_GUC_SUBMISSION(dev_priv))
> > +	if (!USES_GUC_SUBMISSION(i915))
> >   		return -ENODEV;
> >   	for (index = 0; index < GUC_MAX_STAGE_DESCRIPTORS; index++, desc++) {
> > @@ -2066,7 +2066,7 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data)
> >   			   desc->wq_addr, desc->wq_size);
> >   		seq_putc(m, '\n');
> > -		for_each_engine_masked(engine, dev_priv, client->engines, tmp) {
> > +		for_each_engine_masked(engine, i915, client->engines, tmp) {
> >   			u32 guc_engine_id = engine->guc_id;
> >   			struct guc_execlist_context *lrc =
> >   						&desc->lrc[guc_engine_id];
> > @@ -2088,19 +2088,19 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data)
> >   static int i915_guc_log_dump(struct seq_file *m, void *data)
> >   {
> >   	struct drm_info_node *node = m->private;
> > -	struct drm_i915_private *dev_priv = node_to_i915(node);
> > +	struct drm_i915_private *i915 = node_to_i915(node);
> >   	bool dump_load_err = !!node->info_ent->data;
> >   	struct drm_i915_gem_object *obj = NULL;
> >   	u32 *log;
> >   	int i = 0;
> > -	if (!HAS_GUC(dev_priv))
> > +	if (!HAS_GUC(i915))
> >   		return -ENODEV;
> >   	if (dump_load_err)
> > -		obj = dev_priv->guc.load_err_log;
> > -	else if (dev_priv->guc.log.vma)
> > -		obj = dev_priv->guc.log.vma->obj;
> > +		obj = i915->guc.load_err_log;
> > +	else if (i915->guc.log.vma)
> > +		obj = i915->guc.log.vma->obj;
> >   	if (!obj)
> >   		return 0;
> > @@ -2126,24 +2126,24 @@ static int i915_guc_log_dump(struct seq_file *m, void *data)
> >   static int i915_guc_log_level_get(void *data, u64 *val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> > -	if (!USES_GUC(dev_priv))
> > +	if (!USES_GUC(i915))
> >   		return -ENODEV;
> > -	*val = intel_guc_log_get_level(&dev_priv->guc.log);
> > +	*val = intel_guc_log_get_level(&i915->guc.log);
> >   	return 0;
> >   }
> >   static int i915_guc_log_level_set(void *data, u64 val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> > -	if (!USES_GUC(dev_priv))
> > +	if (!USES_GUC(i915))
> >   		return -ENODEV;
> > -	return intel_guc_log_set_level(&dev_priv->guc.log, val);
> > +	return intel_guc_log_set_level(&i915->guc.log, val);
> >   }
> >   DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops,
> > @@ -2152,14 +2152,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops,
> >   static int i915_guc_log_relay_open(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	if (!USES_GUC(dev_priv))
> > +	if (!USES_GUC(i915))
> >   		return -ENODEV;
> > -	file->private_data = &dev_priv->guc.log;
> > +	file->private_data = &i915->guc.log;
> > -	return intel_guc_log_relay_open(&dev_priv->guc.log);
> > +	return intel_guc_log_relay_open(&i915->guc.log);
> >   }
> >   static ssize_t
> > @@ -2177,9 +2177,9 @@ i915_guc_log_relay_write(struct file *filp,
> >   static int i915_guc_log_relay_release(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	intel_guc_log_relay_close(&dev_priv->guc.log);
> > +	intel_guc_log_relay_close(&i915->guc.log);
> >   	return 0;
> >   }
> > @@ -2205,12 +2205,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data)
> >   		"sink internal error",
> >   	};
> >   	struct drm_connector *connector = m->private;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_dp *intel_dp =
> >   		enc_to_intel_dp(&intel_attached_encoder(connector)->base);
> >   	int ret;
> > -	if (!CAN_PSR(dev_priv)) {
> > +	if (!CAN_PSR(i915)) {
> >   		seq_puts(m, "PSR Unsupported\n");
> >   		return -ENODEV;
> >   	}
> > @@ -2236,12 +2236,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data)
> >   DEFINE_SHOW_ATTRIBUTE(i915_psr_sink_status);
> >   static void
> > -psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m)
> > +psr_source_status(struct drm_i915_private *i915, struct seq_file *m)
> >   {
> >   	u32 val, status_val;
> >   	const char *status = "unknown";
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		static const char * const live_status[] = {
> >   			"IDLE",
> >   			"CAPTURE",
> > @@ -2283,14 +2283,14 @@ psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m)
> >   static int i915_edp_psr_status(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct i915_psr *psr = &dev_priv->psr;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct i915_psr *psr = &i915->psr;
> >   	intel_wakeref_t wakeref;
> >   	const char *status;
> >   	bool enabled;
> >   	u32 val;
> > -	if (!HAS_PSR(dev_priv))
> > +	if (!HAS_PSR(i915))
> >   		return -ENODEV;
> >   	seq_printf(m, "Sink support: %s", yesno(psr->sink_support));
> > @@ -2301,7 +2301,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> >   	if (!psr->sink_support)
> >   		return 0;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	mutex_lock(&psr->lock);
> >   	if (psr->enabled)
> > @@ -2322,14 +2322,14 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> >   	}
> >   	seq_printf(m, "Source PSR ctl: %s [0x%08x]\n",
> >   		   enableddisabled(enabled), val);
> > -	psr_source_status(dev_priv, m);
> > +	psr_source_status(i915, m);
> >   	seq_printf(m, "Busy frontbuffer bits: 0x%08x\n",
> >   		   psr->busy_frontbuffer_bits);
> >   	/*
> >   	 * SKL+ Perf counter is reset to 0 everytime DC state is entered
> >   	 */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		val = I915_READ(EDP_PSR_PERF_CNT) & EDP_PSR_PERF_CNT_MASK;
> >   		seq_printf(m, "Performance counter: %u\n", val);
> >   	}
> > @@ -2365,7 +2365,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> >   unlock:
> >   	mutex_unlock(&psr->lock);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> > @@ -2373,20 +2373,20 @@ static int i915_edp_psr_status(struct seq_file *m, void *data)
> >   static int
> >   i915_edp_psr_debug_set(void *data, u64 val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> >   	intel_wakeref_t wakeref;
> >   	int ret;
> > -	if (!CAN_PSR(dev_priv))
> > +	if (!CAN_PSR(i915))
> >   		return -ENODEV;
> >   	DRM_DEBUG_KMS("Setting PSR debug to %llx\n", val);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	ret = intel_psr_debug_set(dev_priv, val);
> > +	ret = intel_psr_debug_set(i915, val);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return ret;
> >   }
> > @@ -2394,12 +2394,12 @@ i915_edp_psr_debug_set(void *data, u64 val)
> >   static int
> >   i915_edp_psr_debug_get(void *data, u64 *val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> > -	if (!CAN_PSR(dev_priv))
> > +	if (!CAN_PSR(i915))
> >   		return -ENODEV;
> > -	*val = READ_ONCE(dev_priv->psr.debug);
> > +	*val = READ_ONCE(i915->psr.debug);
> >   	return 0;
> >   }
> > @@ -2409,19 +2409,19 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_edp_psr_debug_fops,
> >   static int i915_energy_uJ(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	unsigned long long power;
> >   	intel_wakeref_t wakeref;
> >   	u32 units;
> > -	if (INTEL_GEN(dev_priv) < 6)
> > +	if (INTEL_GEN(i915) < 6)
> >   		return -ENODEV;
> >   	if (rdmsrl_safe(MSR_RAPL_POWER_UNIT, &power))
> >   		return -ENODEV;
> >   	units = (power & 0x1f00) >> 8;
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		power = I915_READ(MCH_SECP_NRG_STTS);
> >   	power = (1000000 * power) >> units; /* convert to uJ */
> > @@ -2432,21 +2432,21 @@ static int i915_energy_uJ(struct seq_file *m, void *data)
> >   static int i915_runtime_pm_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > -	if (!HAS_RUNTIME_PM(dev_priv))
> > +	if (!HAS_RUNTIME_PM(i915))
> >   		seq_puts(m, "Runtime power management not supported\n");
> >   	seq_printf(m, "Runtime power status: %s\n",
> > -		   enableddisabled(!dev_priv->power_domains.wakeref));
> > +		   enableddisabled(!i915->power_domains.wakeref));
> > -	seq_printf(m, "GPU idle: %s\n", yesno(!dev_priv->gt.awake));
> > +	seq_printf(m, "GPU idle: %s\n", yesno(!i915->gt.awake));
> >   	seq_printf(m, "IRQs disabled: %s\n",
> > -		   yesno(!intel_irqs_enabled(dev_priv)));
> > +		   yesno(!intel_irqs_enabled(i915)));
> >   #ifdef CONFIG_PM
> >   	seq_printf(m, "Usage count: %d\n",
> > -		   atomic_read(&dev_priv->drm.dev->power.usage_count));
> > +		   atomic_read(&i915->drm.dev->power.usage_count));
> >   #else
> >   	seq_printf(m, "Device Power Management (CONFIG_PM) disabled\n");
> >   #endif
> > @@ -2457,7 +2457,7 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused)
> >   	if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)) {
> >   		struct drm_printer p = drm_seq_file_printer(m);
> > -		print_intel_runtime_pm_wakeref(dev_priv, &p);
> > +		print_intel_runtime_pm_wakeref(i915, &p);
> >   	}
> >   	return 0;
> > @@ -2465,8 +2465,8 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused)
> >   static int i915_power_domain_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	int i;
> >   	mutex_lock(&power_domains->lock);
> > @@ -2493,16 +2493,16 @@ static int i915_power_domain_info(struct seq_file *m, void *unused)
> >   static int i915_dmc_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	intel_wakeref_t wakeref;
> >   	struct intel_csr *csr;
> > -	if (!HAS_CSR(dev_priv))
> > +	if (!HAS_CSR(i915))
> >   		return -ENODEV;
> > -	csr = &dev_priv->csr;
> > +	csr = &i915->csr;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	seq_printf(m, "fw loaded: %s\n", yesno(csr->dmc_payload != NULL));
> >   	seq_printf(m, "path: %s\n", csr->fw_path);
> > @@ -2513,13 +2513,13 @@ static int i915_dmc_info(struct seq_file *m, void *unused)
> >   	seq_printf(m, "version: %d.%d\n", CSR_VERSION_MAJOR(csr->version),
> >   		   CSR_VERSION_MINOR(csr->version));
> > -	if (WARN_ON(INTEL_GEN(dev_priv) > 11))
> > +	if (WARN_ON(INTEL_GEN(i915) > 11))
> >   		goto out;
> >   	seq_printf(m, "DC3 -> DC5 count: %d\n",
> > -		   I915_READ(IS_BROXTON(dev_priv) ? BXT_CSR_DC3_DC5_COUNT :
> > +		   I915_READ(IS_BROXTON(i915) ? BXT_CSR_DC3_DC5_COUNT :
> >   						    SKL_CSR_DC3_DC5_COUNT));
> > -	if (!IS_GEN9_LP(dev_priv))
> > +	if (!IS_GEN9_LP(i915))
> >   		seq_printf(m, "DC5 -> DC6 count: %d\n",
> >   			   I915_READ(SKL_CSR_DC5_DC6_COUNT));
> > @@ -2528,7 +2528,7 @@ static int i915_dmc_info(struct seq_file *m, void *unused)
> >   	seq_printf(m, "ssp base: 0x%08x\n", I915_READ(CSR_SSP_BASE));
> >   	seq_printf(m, "htp: 0x%08x\n", I915_READ(CSR_HTP_SKL));
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> > @@ -2548,8 +2548,8 @@ static void intel_encoder_info(struct seq_file *m,
> >   			       struct intel_crtc *intel_crtc,
> >   			       struct intel_encoder *intel_encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_crtc *crtc = &intel_crtc->base;
> >   	struct intel_connector *intel_connector;
> >   	struct drm_encoder *encoder;
> > @@ -2575,8 +2575,8 @@ static void intel_encoder_info(struct seq_file *m,
> >   static void intel_crtc_info(struct seq_file *m, struct intel_crtc *intel_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_crtc *crtc = &intel_crtc->base;
> >   	struct intel_encoder *intel_encoder;
> >   	struct drm_plane_state *plane_state = crtc->primary->state;
> > @@ -2731,8 +2731,8 @@ static void plane_rotation(char *buf, size_t bufsize, unsigned int rotation)
> >   static void intel_plane_info(struct seq_file *m, struct intel_crtc *intel_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_plane *intel_plane;
> >   	for_each_intel_plane_on_crtc(dev, intel_crtc, intel_plane) {
> > @@ -2805,14 +2805,14 @@ static void intel_scaler_info(struct seq_file *m, struct intel_crtc *intel_crtc)
> >   static int i915_display_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_crtc *crtc;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	intel_wakeref_t wakeref;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	seq_printf(m, "CRTC info\n");
> >   	seq_printf(m, "---------\n");
> > @@ -2861,42 +2861,42 @@ static int i915_display_info(struct seq_file *m, void *unused)
> >   	drm_connector_list_iter_end(&conn_iter);
> >   	mutex_unlock(&dev->mode_config.mutex);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_engine_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	struct intel_engine_cs *engine;
> >   	intel_wakeref_t wakeref;
> >   	enum intel_engine_id id;
> >   	struct drm_printer p;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	seq_printf(m, "GT awake? %s [%d]\n",
> > -		   yesno(dev_priv->gt.awake),
> > -		   atomic_read(&dev_priv->gt.wakeref.count));
> > +		   yesno(i915->gt.awake),
> > +		   atomic_read(&i915->gt.wakeref.count));
> >   	seq_printf(m, "CS timestamp frequency: %u kHz\n",
> > -		   RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz);
> > +		   RUNTIME_INFO(i915)->cs_timestamp_frequency_khz);
> >   	p = drm_seq_file_printer(m);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		intel_engine_dump(engine, &p, "%s\n", engine->name);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return 0;
> >   }
> >   static int i915_rcs_topology(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	struct drm_printer p = drm_seq_file_printer(m);
> > -	intel_device_info_dump_topology(&RUNTIME_INFO(dev_priv)->sseu, &p);
> > +	intel_device_info_dump_topology(&RUNTIME_INFO(i915)->sseu, &p);
> >   	return 0;
> >   }
> > @@ -2913,13 +2913,13 @@ static int i915_shrinker_info(struct seq_file *m, void *unused)
> >   static int i915_shared_dplls_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	int i;
> >   	drm_modeset_lock_all(dev);
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > -		struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> > +		struct intel_shared_dpll *pll = &i915->shared_dplls[i];
> >   		seq_printf(m, "DPLL%i: %s, id: %i\n", i, pll->info->name,
> >   			   pll->info->id);
> > @@ -2977,28 +2977,28 @@ static int i915_wa_registers(struct seq_file *m, void *unused)
> >   static int i915_ipc_status_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	seq_printf(m, "Isochronous Priority Control: %s\n",
> > -			yesno(dev_priv->ipc_enabled));
> > +			yesno(i915->ipc_enabled));
> >   	return 0;
> >   }
> >   static int i915_ipc_status_open(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	if (!HAS_IPC(dev_priv))
> > +	if (!HAS_IPC(i915))
> >   		return -ENODEV;
> > -	return single_open(file, i915_ipc_status_show, dev_priv);
> > +	return single_open(file, i915_ipc_status_show, i915);
> >   }
> >   static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf,
> >   				     size_t len, loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	intel_wakeref_t wakeref;
> >   	bool enable;
> >   	int ret;
> > @@ -3007,12 +3007,12 @@ static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf,
> >   	if (ret < 0)
> >   		return ret;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > -		if (!dev_priv->ipc_enabled && enable)
> > +	with_intel_runtime_pm(i915, wakeref) {
> > +		if (!i915->ipc_enabled && enable)
> >   			DRM_INFO("Enabling IPC: WM will be proper only after next commit\n");
> > -		dev_priv->wm.distrust_bios_wm = true;
> > -		dev_priv->ipc_enabled = enable;
> > -		intel_enable_ipc(dev_priv);
> > +		i915->wm.distrust_bios_wm = true;
> > +		i915->ipc_enabled = enable;
> > +		intel_enable_ipc(i915);
> >   	}
> >   	return len;
> > @@ -3029,19 +3029,19 @@ static const struct file_operations i915_ipc_status_fops = {
> >   static int i915_ddb_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct skl_ddb_entry *entry;
> >   	struct intel_crtc *crtc;
> > -	if (INTEL_GEN(dev_priv) < 9)
> > +	if (INTEL_GEN(i915) < 9)
> >   		return -ENODEV;
> >   	drm_modeset_lock_all(dev);
> >   	seq_printf(m, "%-15s%8s%8s%8s\n", "", "Start", "End", "Size");
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		enum pipe pipe = crtc->pipe;
> > @@ -3070,8 +3070,8 @@ static void drrs_status_per_crtc(struct seq_file *m,
> >   				 struct drm_device *dev,
> >   				 struct intel_crtc *intel_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct i915_drrs *drrs = &dev_priv->drrs;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct i915_drrs *drrs = &i915->drrs;
> >   	int vrefresh = 0;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> > @@ -3085,11 +3085,11 @@ static void drrs_status_per_crtc(struct seq_file *m,
> >   	}
> >   	drm_connector_list_iter_end(&conn_iter);
> > -	if (dev_priv->vbt.drrs_type == STATIC_DRRS_SUPPORT)
> > +	if (i915->vbt.drrs_type == STATIC_DRRS_SUPPORT)
> >   		seq_puts(m, "\tVBT: DRRS_type: Static");
> > -	else if (dev_priv->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT)
> > +	else if (i915->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT)
> >   		seq_puts(m, "\tVBT: DRRS_type: Seamless");
> > -	else if (dev_priv->vbt.drrs_type == DRRS_NOT_SUPPORTED)
> > +	else if (i915->vbt.drrs_type == DRRS_NOT_SUPPORTED)
> >   		seq_puts(m, "\tVBT: DRRS_type: None");
> >   	else
> >   		seq_puts(m, "\tVBT: DRRS_type: FIXME: Unrecognized Value");
> > @@ -3106,7 +3106,7 @@ static void drrs_status_per_crtc(struct seq_file *m,
> >   		/* disable_drrs() will make drrs->dp NULL */
> >   		if (!drrs->dp) {
> >   			seq_puts(m, "Idleness DRRS: Disabled\n");
> > -			if (dev_priv->psr.enabled)
> > +			if (i915->psr.enabled)
> >   				seq_puts(m,
> >   				"\tAs PSR is enabled, DRRS is not enabled\n");
> >   			mutex_unlock(&drrs->mutex);
> > @@ -3143,8 +3143,8 @@ static void drrs_status_per_crtc(struct seq_file *m,
> >   static int i915_drrs_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_crtc *intel_crtc;
> >   	int active_crtc_cnt = 0;
> > @@ -3167,8 +3167,8 @@ static int i915_drrs_status(struct seq_file *m, void *unused)
> >   static int i915_dp_mst_info(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_encoder *intel_encoder;
> >   	struct intel_digital_port *intel_dig_port;
> >   	struct drm_connector *connector;
> > @@ -3257,8 +3257,8 @@ static ssize_t i915_displayport_test_active_write(struct file *file,
> >   static int i915_displayport_test_active_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	struct intel_dp *intel_dp;
> > @@ -3307,8 +3307,8 @@ static const struct file_operations i915_displayport_test_active_fops = {
> >   static int i915_displayport_test_data_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	struct intel_dp *intel_dp;
> > @@ -3351,8 +3351,8 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_data);
> >   static int i915_displayport_test_type_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	struct intel_dp *intel_dp;
> > @@ -3383,19 +3383,19 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_type);
> >   static void wm_latency_show(struct seq_file *m, const u16 wm[8])
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct drm_device *dev = &i915->drm;
> >   	int level;
> >   	int num_levels;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		num_levels = 3;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		num_levels = 1;
> > -	else if (IS_G4X(dev_priv))
> > +	else if (IS_G4X(i915))
> >   		num_levels = 3;
> >   	else
> > -		num_levels = ilk_wm_max_level(dev_priv) + 1;
> > +		num_levels = ilk_wm_max_level(i915) + 1;
> >   	drm_modeset_lock_all(dev);
> > @@ -3406,10 +3406,10 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8])
> >   		 * - WM1+ latency values in 0.5us units
> >   		 * - latencies are in us on gen9/vlv/chv
> >   		 */
> > -		if (INTEL_GEN(dev_priv) >= 9 ||
> > -		    IS_VALLEYVIEW(dev_priv) ||
> > -		    IS_CHERRYVIEW(dev_priv) ||
> > -		    IS_G4X(dev_priv))
> > +		if (INTEL_GEN(i915) >= 9 ||
> > +		    IS_VALLEYVIEW(i915) ||
> > +		    IS_CHERRYVIEW(i915) ||
> > +		    IS_G4X(i915))
> >   			latency *= 10;
> >   		else if (level > 0)
> >   			latency *= 5;
> > @@ -3423,13 +3423,13 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8])
> >   static int pri_wm_latency_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	const u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.pri_latency;
> > +		latencies = i915->wm.pri_latency;
> >   	wm_latency_show(m, latencies);
> > @@ -3438,13 +3438,13 @@ static int pri_wm_latency_show(struct seq_file *m, void *data)
> >   static int spr_wm_latency_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	const u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.spr_latency;
> > +		latencies = i915->wm.spr_latency;
> >   	wm_latency_show(m, latencies);
> > @@ -3453,13 +3453,13 @@ static int spr_wm_latency_show(struct seq_file *m, void *data)
> >   static int cur_wm_latency_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	const u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.cur_latency;
> > +		latencies = i915->wm.cur_latency;
> >   	wm_latency_show(m, latencies);
> > @@ -3468,54 +3468,54 @@ static int cur_wm_latency_show(struct seq_file *m, void *data)
> >   static int pri_wm_latency_open(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> >   		return -ENODEV;
> > -	return single_open(file, pri_wm_latency_show, dev_priv);
> > +	return single_open(file, pri_wm_latency_show, i915);
> >   }
> >   static int spr_wm_latency_open(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		return -ENODEV;
> > -	return single_open(file, spr_wm_latency_show, dev_priv);
> > +	return single_open(file, spr_wm_latency_show, i915);
> >   }
> >   static int cur_wm_latency_open(struct inode *inode, struct file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = inode->i_private;
> > +	struct drm_i915_private *i915 = inode->i_private;
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		return -ENODEV;
> > -	return single_open(file, cur_wm_latency_show, dev_priv);
> > +	return single_open(file, cur_wm_latency_show, i915);
> >   }
> >   static ssize_t wm_latency_write(struct file *file, const char __user *ubuf,
> >   				size_t len, loff_t *offp, u16 wm[8])
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct drm_device *dev = &i915->drm;
> >   	u16 new[8] = { 0 };
> >   	int num_levels;
> >   	int level;
> >   	int ret;
> >   	char tmp[32];
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		num_levels = 3;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		num_levels = 1;
> > -	else if (IS_G4X(dev_priv))
> > +	else if (IS_G4X(i915))
> >   		num_levels = 3;
> >   	else
> > -		num_levels = ilk_wm_max_level(dev_priv) + 1;
> > +		num_levels = ilk_wm_max_level(i915) + 1;
> >   	if (len >= sizeof(tmp))
> >   		return -EINVAL;
> > @@ -3546,13 +3546,13 @@ static ssize_t pri_wm_latency_write(struct file *file, const char __user *ubuf,
> >   				    size_t len, loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.pri_latency;
> > +		latencies = i915->wm.pri_latency;
> >   	return wm_latency_write(file, ubuf, len, offp, latencies);
> >   }
> > @@ -3561,13 +3561,13 @@ static ssize_t spr_wm_latency_write(struct file *file, const char __user *ubuf,
> >   				    size_t len, loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.spr_latency;
> > +		latencies = i915->wm.spr_latency;
> >   	return wm_latency_write(file, ubuf, len, offp, latencies);
> >   }
> > @@ -3576,13 +3576,13 @@ static ssize_t cur_wm_latency_write(struct file *file, const char __user *ubuf,
> >   				    size_t len, loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	u16 *latencies;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		latencies = dev_priv->wm.skl_latency;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		latencies = i915->wm.skl_latency;
> >   	else
> > -		latencies = dev_priv->wm.cur_latency;
> > +		latencies = i915->wm.cur_latency;
> >   	return wm_latency_write(file, ubuf, len, offp, latencies);
> >   }
> > @@ -3752,14 +3752,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_drop_caches_fops,
> >   static int
> >   i915_cache_sharing_get(void *data, u64 *val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> >   	intel_wakeref_t wakeref;
> >   	u32 snpcr = 0;
> > -	if (!(IS_GEN_RANGE(dev_priv, 6, 7)))
> > +	if (!(IS_GEN_RANGE(i915, 6, 7)))
> >   		return -ENODEV;
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		snpcr = I915_READ(GEN6_MBCUNIT_SNPCR);
> >   	*val = (snpcr & GEN6_MBC_SNPCR_MASK) >> GEN6_MBC_SNPCR_SHIFT;
> > @@ -3770,17 +3770,17 @@ i915_cache_sharing_get(void *data, u64 *val)
> >   static int
> >   i915_cache_sharing_set(void *data, u64 val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> >   	intel_wakeref_t wakeref;
> > -	if (!(IS_GEN_RANGE(dev_priv, 6, 7)))
> > +	if (!(IS_GEN_RANGE(i915, 6, 7)))
> >   		return -ENODEV;
> >   	if (val > 3)
> >   		return -EINVAL;
> >   	DRM_DEBUG_DRIVER("Manually setting uncore sharing to %llu\n", val);
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		u32 snpcr;
> >   		/* Update the cache sharing policy here as well */
> > @@ -3797,7 +3797,7 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_cache_sharing_fops,
> >   			i915_cache_sharing_get, i915_cache_sharing_set,
> >   			"%llu\n");
> > -static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv,
> > +static void cherryview_sseu_device_status(struct drm_i915_private *i915,
> >   					  struct sseu_dev_info *sseu)
> >   {
> >   #define SS_MAX 2
> > @@ -3830,11 +3830,11 @@ static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv,
> >   #undef SS_MAX
> >   }
> > -static void gen10_sseu_device_status(struct drm_i915_private *dev_priv,
> > +static void gen10_sseu_device_status(struct drm_i915_private *i915,
> >   				     struct sseu_dev_info *sseu)
> >   {
> >   #define SS_MAX 6
> > -	const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv);
> > +	const struct intel_runtime_info *info = RUNTIME_INFO(i915);
> >   	u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2];
> >   	int s, ss;
> > @@ -3886,11 +3886,11 @@ static void gen10_sseu_device_status(struct drm_i915_private *dev_priv,
> >   #undef SS_MAX
> >   }
> > -static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
> > +static void gen9_sseu_device_status(struct drm_i915_private *i915,
> >   				    struct sseu_dev_info *sseu)
> >   {
> >   #define SS_MAX 3
> > -	const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv);
> > +	const struct intel_runtime_info *info = RUNTIME_INFO(i915);
> >   	u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2];
> >   	int s, ss;
> > @@ -3916,14 +3916,14 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
> >   		sseu->slice_mask |= BIT(s);
> > -		if (IS_GEN9_BC(dev_priv))
> > +		if (IS_GEN9_BC(i915))
> >   			sseu->subslice_mask[s] =
> > -				RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s];
> > +				RUNTIME_INFO(i915)->sseu.subslice_mask[s];
> >   		for (ss = 0; ss < info->sseu.max_subslices; ss++) {
> >   			unsigned int eu_cnt;
> > -			if (IS_GEN9_LP(dev_priv)) {
> > +			if (IS_GEN9_LP(i915)) {
> >   				if (!(s_reg[s] & (GEN9_PGCTL_SS_ACK(ss))))
> >   					/* skip disabled subslice */
> >   					continue;
> > @@ -3942,7 +3942,7 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv,
> >   #undef SS_MAX
> >   }
> > -static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> > +static void broadwell_sseu_device_status(struct drm_i915_private *i915,
> >   					 struct sseu_dev_info *sseu)
> >   {
> >   	u32 slice_info = I915_READ(GEN8_GT_SLICE_INFO);
> > @@ -3952,10 +3952,10 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> >   	if (sseu->slice_mask) {
> >   		sseu->eu_per_subslice =
> > -			RUNTIME_INFO(dev_priv)->sseu.eu_per_subslice;
> > +			RUNTIME_INFO(i915)->sseu.eu_per_subslice;
> >   		for (s = 0; s < fls(sseu->slice_mask); s++) {
> >   			sseu->subslice_mask[s] =
> > -				RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s];
> > +				RUNTIME_INFO(i915)->sseu.subslice_mask[s];
> >   		}
> >   		sseu->eu_total = sseu->eu_per_subslice *
> >   				 intel_sseu_subslice_total(sseu);
> > @@ -3963,7 +3963,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> >   		/* subtract fused off EU(s) from enabled slice(s) */
> >   		for (s = 0; s < fls(sseu->slice_mask); s++) {
> >   			u8 subslice_7eu =
> > -				RUNTIME_INFO(dev_priv)->sseu.subslice_7eu[s];
> > +				RUNTIME_INFO(i915)->sseu.subslice_7eu[s];
> >   			sseu->eu_total -= hweight8(subslice_7eu);
> >   		}
> > @@ -3973,7 +3973,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv,
> >   static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
> >   				 const struct sseu_dev_info *sseu)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	const char *type = is_available_info ? "Available" : "Enabled";
> >   	int s;
> > @@ -3995,8 +3995,8 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
> >   	if (!is_available_info)
> >   		return;
> > -	seq_printf(m, "  Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(dev_priv)));
> > -	if (HAS_POOLED_EU(dev_priv))
> > +	seq_printf(m, "  Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(i915)));
> > +	if (HAS_POOLED_EU(i915))
> >   		seq_printf(m, "  Min EU in pool: %u\n", sseu->min_eu_in_pool);
> >   	seq_printf(m, "  Has Slice Power Gating: %s\n",
> > @@ -4009,32 +4009,32 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info,
> >   static int i915_sseu_status(struct seq_file *m, void *unused)
> >   {
> > -	struct drm_i915_private *dev_priv = node_to_i915(m->private);
> > +	struct drm_i915_private *i915 = node_to_i915(m->private);
> >   	struct sseu_dev_info sseu;
> >   	intel_wakeref_t wakeref;
> > -	if (INTEL_GEN(dev_priv) < 8)
> > +	if (INTEL_GEN(i915) < 8)
> >   		return -ENODEV;
> >   	seq_puts(m, "SSEU Device Info\n");
> > -	i915_print_sseu_info(m, true, &RUNTIME_INFO(dev_priv)->sseu);
> > +	i915_print_sseu_info(m, true, &RUNTIME_INFO(i915)->sseu);
> >   	seq_puts(m, "SSEU Device Status\n");
> >   	memset(&sseu, 0, sizeof(sseu));
> > -	sseu.max_slices = RUNTIME_INFO(dev_priv)->sseu.max_slices;
> > -	sseu.max_subslices = RUNTIME_INFO(dev_priv)->sseu.max_subslices;
> > +	sseu.max_slices = RUNTIME_INFO(i915)->sseu.max_slices;
> > +	sseu.max_subslices = RUNTIME_INFO(i915)->sseu.max_subslices;
> >   	sseu.max_eus_per_subslice =
> > -		RUNTIME_INFO(dev_priv)->sseu.max_eus_per_subslice;
> > -
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > -		if (IS_CHERRYVIEW(dev_priv))
> > -			cherryview_sseu_device_status(dev_priv, &sseu);
> > -		else if (IS_BROADWELL(dev_priv))
> > -			broadwell_sseu_device_status(dev_priv, &sseu);
> > -		else if (IS_GEN(dev_priv, 9))
> > -			gen9_sseu_device_status(dev_priv, &sseu);
> > -		else if (INTEL_GEN(dev_priv) >= 10)
> > -			gen10_sseu_device_status(dev_priv, &sseu);
> > +		RUNTIME_INFO(i915)->sseu.max_eus_per_subslice;
> > +
> > +	with_intel_runtime_pm(i915, wakeref) {
> > +		if (IS_CHERRYVIEW(i915))
> > +			cherryview_sseu_device_status(i915, &sseu);
> > +		else if (IS_BROADWELL(i915))
> > +			broadwell_sseu_device_status(i915, &sseu);
> > +		else if (IS_GEN(i915, 9))
> > +			gen9_sseu_device_status(i915, &sseu);
> > +		else if (INTEL_GEN(i915) >= 10)
> > +			gen10_sseu_device_status(i915, &sseu);
> >   	}
> >   	i915_print_sseu_info(m, false, &sseu);
> > @@ -4077,15 +4077,15 @@ static const struct file_operations i915_forcewake_fops = {
> >   static int i915_hpd_storm_ctl_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct i915_hotplug *hotplug = &dev_priv->hotplug;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct i915_hotplug *hotplug = &i915->hotplug;
> >   	/* Synchronize with everything first in case there's been an HPD
> >   	 * storm, but we haven't finished handling it in the kernel yet
> >   	 */
> > -	synchronize_irq(dev_priv->drm.irq);
> > -	flush_work(&dev_priv->hotplug.dig_port_work);
> > -	flush_work(&dev_priv->hotplug.hotplug_work);
> > +	synchronize_irq(i915->drm.irq);
> > +	flush_work(&i915->hotplug.dig_port_work);
> > +	flush_work(&i915->hotplug.hotplug_work);
> >   	seq_printf(m, "Threshold: %d\n", hotplug->hpd_storm_threshold);
> >   	seq_printf(m, "Detected: %s\n",
> > @@ -4099,8 +4099,8 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file,
> >   					loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct i915_hotplug *hotplug = &dev_priv->hotplug;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct i915_hotplug *hotplug = &i915->hotplug;
> >   	unsigned int new_threshold;
> >   	int i;
> >   	char *newline;
> > @@ -4130,15 +4130,15 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file,
> >   	else
> >   		DRM_DEBUG_KMS("Disabling HPD storm detection\n");
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	hotplug->hpd_storm_threshold = new_threshold;
> >   	/* Reset the HPD storm stats so we don't accidentally trigger a storm */
> >   	for_each_hpd_pin(i)
> >   		hotplug->stats[i].count = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/* Re-enable hpd immediately if we were in an irq storm */
> > -	flush_delayed_work(&dev_priv->hotplug.reenable_work);
> > +	flush_delayed_work(&i915->hotplug.reenable_work);
> >   	return len;
> >   }
> > @@ -4159,10 +4159,10 @@ static const struct file_operations i915_hpd_storm_ctl_fops = {
> >   static int i915_hpd_short_storm_ctl_show(struct seq_file *m, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = m->private;
> > +	struct drm_i915_private *i915 = m->private;
> >   	seq_printf(m, "Enabled: %s\n",
> > -		   yesno(dev_priv->hotplug.hpd_short_storm_enabled));
> > +		   yesno(i915->hotplug.hpd_short_storm_enabled));
> >   	return 0;
> >   }
> > @@ -4179,8 +4179,8 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file,
> >   					      size_t len, loff_t *offp)
> >   {
> >   	struct seq_file *m = file->private_data;
> > -	struct drm_i915_private *dev_priv = m->private;
> > -	struct i915_hotplug *hotplug = &dev_priv->hotplug;
> > +	struct drm_i915_private *i915 = m->private;
> > +	struct i915_hotplug *hotplug = &i915->hotplug;
> >   	char *newline;
> >   	char tmp[16];
> >   	int i;
> > @@ -4201,22 +4201,22 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file,
> >   	/* Reset to the "default" state for this system */
> >   	if (strcmp(tmp, "reset") == 0)
> > -		new_state = !HAS_DP_MST(dev_priv);
> > +		new_state = !HAS_DP_MST(i915);
> >   	else if (kstrtobool(tmp, &new_state) != 0)
> >   		return -EINVAL;
> >   	DRM_DEBUG_KMS("%sabling HPD short storm detection\n",
> >   		      new_state ? "En" : "Dis");
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	hotplug->hpd_short_storm_enabled = new_state;
> >   	/* Reset the HPD storm stats so we don't accidentally trigger a storm */
> >   	for_each_hpd_pin(i)
> >   		hotplug->stats[i].count = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/* Re-enable hpd immediately if we were in an irq storm */
> > -	flush_delayed_work(&dev_priv->hotplug.reenable_work);
> > +	flush_delayed_work(&i915->hotplug.reenable_work);
> >   	return len;
> >   }
> > @@ -4232,11 +4232,11 @@ static const struct file_operations i915_hpd_short_storm_ctl_fops = {
> >   static int i915_drrs_ctl_set(void *data, u64 val)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = data;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_crtc *crtc;
> > -	if (INTEL_GEN(dev_priv) < 7)
> > +	if (INTEL_GEN(i915) < 7)
> >   		return -ENODEV;
> >   	for_each_intel_crtc(dev, crtc) {
> > @@ -4305,9 +4305,9 @@ i915_fifo_underrun_reset_write(struct file *filp,
> >   			       const char __user *ubuf,
> >   			       size_t cnt, loff_t *ppos)
> >   {
> > -	struct drm_i915_private *dev_priv = filp->private_data;
> > +	struct drm_i915_private *i915 = filp->private_data;
> >   	struct intel_crtc *intel_crtc;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	int ret;
> >   	bool reset;
> > @@ -4347,7 +4347,7 @@ i915_fifo_underrun_reset_write(struct file *filp,
> >   			return ret;
> >   	}
> > -	ret = intel_fbc_reset_underrun(dev_priv);
> > +	ret = intel_fbc_reset_underrun(i915);
> >   	if (ret)
> >   		return ret;
> > @@ -4437,9 +4437,9 @@ static const struct i915_debugfs_files {
> >   	{"i915_edp_psr_debug", &i915_edp_psr_debug_fops}
> >   };
> > -int i915_debugfs_register(struct drm_i915_private *dev_priv)
> > +int i915_debugfs_register(struct drm_i915_private *i915)
> >   {
> > -	struct drm_minor *minor = dev_priv->drm.primary;
> > +	struct drm_minor *minor = i915->drm.primary;
> >   	struct dentry *ent;
> >   	int i;
> > @@ -4692,7 +4692,7 @@ static const struct file_operations i915_dsc_fec_support_fops = {
> >   int i915_debugfs_connector_add(struct drm_connector *connector)
> >   {
> >   	struct dentry *root = connector->debugfs_entry;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	/* The connector must have been registered beforehands. */
> >   	if (!root)
> > @@ -4717,7 +4717,7 @@ int i915_debugfs_connector_add(struct drm_connector *connector)
> >   				    connector, &i915_hdcp_sink_capability_fops);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 10 &&
> > +	if (INTEL_GEN(i915) >= 10 &&
> >   	    (connector->connector_type == DRM_MODE_CONNECTOR_DisplayPort ||
> >   	     connector->connector_type == DRM_MODE_CONNECTOR_eDP))
> >   		debugfs_create_file("i915_dsc_fec_support", S_IRUGO, root,
> > diff --git a/drivers/gpu/drm/i915/i915_debugfs.h b/drivers/gpu/drm/i915/i915_debugfs.h
> > index c0cd22eb916d..3e99eaf8de02 100644
> > --- a/drivers/gpu/drm/i915/i915_debugfs.h
> > +++ b/drivers/gpu/drm/i915/i915_debugfs.h
> > @@ -10,10 +10,10 @@ struct drm_i915_private;
> >   struct drm_connector;
> >   #ifdef CONFIG_DEBUG_FS
> > -int i915_debugfs_register(struct drm_i915_private *dev_priv);
> > +int i915_debugfs_register(struct drm_i915_private *i915);
> >   int i915_debugfs_connector_add(struct drm_connector *connector);
> >   #else
> > -static inline int i915_debugfs_register(struct drm_i915_private *dev_priv) { return 0; }
> > +static inline int i915_debugfs_register(struct drm_i915_private *i915) { return 0; }
> >   static inline int i915_debugfs_connector_add(struct drm_connector *connector) { return 0; }
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c
> > index 81ff2c78fd55..a09d028ec8d8 100644
> > --- a/drivers/gpu/drm/i915/i915_drv.c
> > +++ b/drivers/gpu/drm/i915/i915_drv.c
> > @@ -108,11 +108,11 @@ bool i915_error_injected(void)
> >   		    "providing the dmesg log by booting with drm.debug=0xf"
> >   void
> > -__i915_printk(struct drm_i915_private *dev_priv, const char *level,
> > +__i915_printk(struct drm_i915_private *i915, const char *level,
> >   	      const char *fmt, ...)
> >   {
> >   	static bool shown_bug_once;
> > -	struct device *kdev = dev_priv->drm.dev;
> > +	struct device *kdev = i915->drm.dev;
> >   	bool is_error = level[1] <= KERN_ERR[1];
> >   	bool is_debug = level[1] == KERN_DEBUG[1];
> >   	struct va_format vaf;
> > @@ -148,74 +148,74 @@ __i915_printk(struct drm_i915_private *dev_priv, const char *level,
> >   /* Map PCH device id to PCH type, or PCH_NONE if unknown. */
> >   static enum intel_pch
> > -intel_pch_type(const struct drm_i915_private *dev_priv, unsigned short id)
> > +intel_pch_type(const struct drm_i915_private *i915, unsigned short id)
> >   {
> >   	switch (id) {
> >   	case INTEL_PCH_IBX_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Ibex Peak PCH\n");
> > -		WARN_ON(!IS_GEN(dev_priv, 5));
> > +		WARN_ON(!IS_GEN(i915, 5));
> >   		return PCH_IBX;
> >   	case INTEL_PCH_CPT_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found CougarPoint PCH\n");
> > -		WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv));
> > +		WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915));
> >   		return PCH_CPT;
> >   	case INTEL_PCH_PPT_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found PantherPoint PCH\n");
> > -		WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv));
> > +		WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915));
> >   		/* PantherPoint is CPT compatible */
> >   		return PCH_CPT;
> >   	case INTEL_PCH_LPT_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found LynxPoint PCH\n");
> > -		WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> > -		WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv));
> > +		WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> > +		WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915));
> >   		return PCH_LPT;
> >   	case INTEL_PCH_LPT_LP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found LynxPoint LP PCH\n");
> > -		WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> > -		WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv));
> > +		WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> > +		WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915));
> >   		return PCH_LPT;
> >   	case INTEL_PCH_WPT_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found WildcatPoint PCH\n");
> > -		WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> > -		WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv));
> > +		WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> > +		WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915));
> >   		/* WildcatPoint is LPT compatible */
> >   		return PCH_LPT;
> >   	case INTEL_PCH_WPT_LP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found WildcatPoint LP PCH\n");
> > -		WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv));
> > -		WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv));
> > +		WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915));
> > +		WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915));
> >   		/* WildcatPoint is LPT compatible */
> >   		return PCH_LPT;
> >   	case INTEL_PCH_SPT_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found SunrisePoint PCH\n");
> > -		WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv));
> > +		WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915));
> >   		return PCH_SPT;
> >   	case INTEL_PCH_SPT_LP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found SunrisePoint LP PCH\n");
> > -		WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv));
> > +		WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915));
> >   		return PCH_SPT;
> >   	case INTEL_PCH_KBP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Kaby Lake PCH (KBP)\n");
> > -		WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv) &&
> > -			!IS_COFFEELAKE(dev_priv));
> > +		WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915) &&
> > +			!IS_COFFEELAKE(i915));
> >   		/* KBP is SPT compatible */
> >   		return PCH_SPT;
> >   	case INTEL_PCH_CNP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Cannon Lake PCH (CNP)\n");
> > -		WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv));
> > +		WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915));
> >   		return PCH_CNP;
> >   	case INTEL_PCH_CNP_LP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Cannon Lake LP PCH (CNP-LP)\n");
> > -		WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv));
> > +		WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915));
> >   		return PCH_CNP;
> >   	case INTEL_PCH_CMP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Comet Lake PCH (CMP)\n");
> > -		WARN_ON(!IS_COFFEELAKE(dev_priv));
> > +		WARN_ON(!IS_COFFEELAKE(i915));
> >   		/* CometPoint is CNP Compatible */
> >   		return PCH_CNP;
> >   	case INTEL_PCH_ICP_DEVICE_ID_TYPE:
> >   		DRM_DEBUG_KMS("Found Ice Lake PCH\n");
> > -		WARN_ON(!IS_ICELAKE(dev_priv));
> > +		WARN_ON(!IS_ICELAKE(i915));
> >   		return PCH_ICP;
> >   	default:
> >   		return PCH_NONE;
> > @@ -233,7 +233,7 @@ static bool intel_is_virt_pch(unsigned short id,
> >   }
> >   static unsigned short
> > -intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> > +intel_virt_detect_pch(const struct drm_i915_private *i915)
> >   {
> >   	unsigned short id = 0;
> > @@ -244,19 +244,19 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> >   	 * make an educated guess as to which PCH is really there.
> >   	 */
> > -	if (IS_ICELAKE(dev_priv))
> > +	if (IS_ICELAKE(i915))
> >   		id = INTEL_PCH_ICP_DEVICE_ID_TYPE;
> > -	else if (IS_CANNONLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915) || IS_COFFEELAKE(i915))
> >   		id = INTEL_PCH_CNP_DEVICE_ID_TYPE;
> > -	else if (IS_KABYLAKE(dev_priv) || IS_SKYLAKE(dev_priv))
> > +	else if (IS_KABYLAKE(i915) || IS_SKYLAKE(i915))
> >   		id = INTEL_PCH_SPT_DEVICE_ID_TYPE;
> > -	else if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv))
> > +	else if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915))
> >   		id = INTEL_PCH_LPT_LP_DEVICE_ID_TYPE;
> > -	else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		id = INTEL_PCH_LPT_DEVICE_ID_TYPE;
> > -	else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> > +	else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> >   		id = INTEL_PCH_CPT_DEVICE_ID_TYPE;
> > -	else if (IS_GEN(dev_priv, 5))
> > +	else if (IS_GEN(i915, 5))
> >   		id = INTEL_PCH_IBX_DEVICE_ID_TYPE;
> >   	if (id)
> > @@ -267,7 +267,7 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv)
> >   	return id;
> >   }
> > -static void intel_detect_pch(struct drm_i915_private *dev_priv)
> > +static void intel_detect_pch(struct drm_i915_private *i915)
> >   {
> >   	struct pci_dev *pch = NULL;
> > @@ -291,22 +291,22 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
> >   		id = pch->device & INTEL_PCH_DEVICE_ID_MASK;
> > -		pch_type = intel_pch_type(dev_priv, id);
> > +		pch_type = intel_pch_type(i915, id);
> >   		if (pch_type != PCH_NONE) {
> > -			dev_priv->pch_type = pch_type;
> > -			dev_priv->pch_id = id;
> > +			i915->pch_type = pch_type;
> > +			i915->pch_id = id;
> >   			break;
> >   		} else if (intel_is_virt_pch(id, pch->subsystem_vendor,
> >   					 pch->subsystem_device)) {
> > -			id = intel_virt_detect_pch(dev_priv);
> > -			pch_type = intel_pch_type(dev_priv, id);
> > +			id = intel_virt_detect_pch(i915);
> > +			pch_type = intel_pch_type(i915, id);
> >   			/* Sanity check virtual PCH id */
> >   			if (WARN_ON(id && pch_type == PCH_NONE))
> >   				id = 0;
> > -			dev_priv->pch_type = pch_type;
> > -			dev_priv->pch_id = id;
> > +			i915->pch_type = pch_type;
> > +			i915->pch_id = id;
> >   			break;
> >   		}
> >   	}
> > @@ -315,10 +315,10 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
> >   	 * Use PCH_NOP (PCH but no South Display) for PCH platforms without
> >   	 * display.
> >   	 */
> > -	if (pch && !HAS_DISPLAY(dev_priv)) {
> > +	if (pch && !HAS_DISPLAY(i915)) {
> >   		DRM_DEBUG_KMS("Display disabled, reverting to NOP PCH\n");
> > -		dev_priv->pch_type = PCH_NOP;
> > -		dev_priv->pch_id = 0;
> > +		i915->pch_type = PCH_NOP;
> > +		i915->pch_id = 0;
> >   	}
> >   	if (!pch)
> > @@ -330,9 +330,9 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv)
> >   static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   			       struct drm_file *file_priv)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > -	const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > +	const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	drm_i915_getparam_t *param = data;
> >   	int value;
> > @@ -350,40 +350,40 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   		value = pdev->revision;
> >   		break;
> >   	case I915_PARAM_NUM_FENCES_AVAIL:
> > -		value = dev_priv->num_fence_regs;
> > +		value = i915->num_fence_regs;
> >   		break;
> >   	case I915_PARAM_HAS_OVERLAY:
> > -		value = dev_priv->overlay ? 1 : 0;
> > +		value = i915->overlay ? 1 : 0;
> >   		break;
> >   	case I915_PARAM_HAS_BSD:
> > -		value = !!dev_priv->engine[VCS0];
> > +		value = !!i915->engine[VCS0];
> >   		break;
> >   	case I915_PARAM_HAS_BLT:
> > -		value = !!dev_priv->engine[BCS0];
> > +		value = !!i915->engine[BCS0];
> >   		break;
> >   	case I915_PARAM_HAS_VEBOX:
> > -		value = !!dev_priv->engine[VECS0];
> > +		value = !!i915->engine[VECS0];
> >   		break;
> >   	case I915_PARAM_HAS_BSD2:
> > -		value = !!dev_priv->engine[VCS1];
> > +		value = !!i915->engine[VCS1];
> >   		break;
> >   	case I915_PARAM_HAS_LLC:
> > -		value = HAS_LLC(dev_priv);
> > +		value = HAS_LLC(i915);
> >   		break;
> >   	case I915_PARAM_HAS_WT:
> > -		value = HAS_WT(dev_priv);
> > +		value = HAS_WT(i915);
> >   		break;
> >   	case I915_PARAM_HAS_ALIASING_PPGTT:
> > -		value = INTEL_PPGTT(dev_priv);
> > +		value = INTEL_PPGTT(i915);
> >   		break;
> >   	case I915_PARAM_HAS_SEMAPHORES:
> > -		value = !!(dev_priv->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES);
> > +		value = !!(i915->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES);
> >   		break;
> >   	case I915_PARAM_HAS_SECURE_BATCHES:
> >   		value = capable(CAP_SYS_ADMIN);
> >   		break;
> >   	case I915_PARAM_CMD_PARSER_VERSION:
> > -		value = i915_cmd_parser_get_version(dev_priv);
> > +		value = i915_cmd_parser_get_version(i915);
> >   		break;
> >   	case I915_PARAM_SUBSLICE_TOTAL:
> >   		value = intel_sseu_subslice_total(sseu);
> > @@ -397,21 +397,21 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   		break;
> >   	case I915_PARAM_HAS_GPU_RESET:
> >   		value = i915_modparams.enable_hangcheck &&
> > -			intel_has_gpu_reset(dev_priv);
> > -		if (value && intel_has_reset_engine(dev_priv))
> > +			intel_has_gpu_reset(i915);
> > +		if (value && intel_has_reset_engine(i915))
> >   			value = 2;
> >   		break;
> >   	case I915_PARAM_HAS_RESOURCE_STREAMER:
> >   		value = 0;
> >   		break;
> >   	case I915_PARAM_HAS_POOLED_EU:
> > -		value = HAS_POOLED_EU(dev_priv);
> > +		value = HAS_POOLED_EU(i915);
> >   		break;
> >   	case I915_PARAM_MIN_EU_IN_POOL:
> >   		value = sseu->min_eu_in_pool;
> >   		break;
> >   	case I915_PARAM_HUC_STATUS:
> > -		value = intel_huc_check_status(&dev_priv->huc);
> > +		value = intel_huc_check_status(&i915->huc);
> >   		if (value < 0)
> >   			return value;
> >   		break;
> > @@ -423,7 +423,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   		value = i915_gem_mmap_gtt_version();
> >   		break;
> >   	case I915_PARAM_HAS_SCHEDULER:
> > -		value = dev_priv->caps.scheduler;
> > +		value = i915->caps.scheduler;
> >   		break;
> >   	case I915_PARAM_MMAP_VERSION:
> > @@ -456,7 +456,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   		value = 1;
> >   		break;
> >   	case I915_PARAM_HAS_CONTEXT_ISOLATION:
> > -		value = intel_engines_has_context_isolation(dev_priv);
> > +		value = intel_engines_has_context_isolation(i915);
> >   		break;
> >   	case I915_PARAM_SLICE_MASK:
> >   		value = sseu->slice_mask;
> > @@ -469,10 +469,10 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   			return -ENODEV;
> >   		break;
> >   	case I915_PARAM_CS_TIMESTAMP_FREQUENCY:
> > -		value = 1000 * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz;
> > +		value = 1000 * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz;
> >   		break;
> >   	case I915_PARAM_MMAP_GTT_COHERENT:
> > -		value = INTEL_INFO(dev_priv)->has_coherent_ggtt;
> > +		value = INTEL_INFO(i915)->has_coherent_ggtt;
> >   		break;
> >   	default:
> >   		DRM_DEBUG("Unknown parameter %d\n", param->param);
> > @@ -485,13 +485,13 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data,
> >   	return 0;
> >   }
> > -static int i915_get_bridge_dev(struct drm_i915_private *dev_priv)
> > +static int i915_get_bridge_dev(struct drm_i915_private *i915)
> >   {
> > -	int domain = pci_domain_nr(dev_priv->drm.pdev->bus);
> > +	int domain = pci_domain_nr(i915->drm.pdev->bus);
> > -	dev_priv->bridge_dev =
> > +	i915->bridge_dev =
> >   		pci_get_domain_bus_and_slot(domain, 0, PCI_DEVFN(0, 0));
> > -	if (!dev_priv->bridge_dev) {
> > +	if (!i915->bridge_dev) {
> >   		DRM_ERROR("bridge device not found\n");
> >   		return -1;
> >   	}
> > @@ -500,16 +500,16 @@ static int i915_get_bridge_dev(struct drm_i915_private *dev_priv)
> >   /* Allocate space for the MCH regs if needed, return nonzero on error */
> >   static int
> > -intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv)
> > +intel_alloc_mchbar_resource(struct drm_i915_private *i915)
> >   {
> > -	int reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> > +	int reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> >   	u32 temp_lo, temp_hi = 0;
> >   	u64 mchbar_addr;
> >   	int ret;
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > -		pci_read_config_dword(dev_priv->bridge_dev, reg + 4, &temp_hi);
> > -	pci_read_config_dword(dev_priv->bridge_dev, reg, &temp_lo);
> > +	if (INTEL_GEN(i915) >= 4)
> > +		pci_read_config_dword(i915->bridge_dev, reg + 4, &temp_hi);
> > +	pci_read_config_dword(i915->bridge_dev, reg, &temp_lo);
> >   	mchbar_addr = ((u64)temp_hi << 32) | temp_lo;
> >   	/* If ACPI doesn't have it, assume we need to allocate it ourselves */
> > @@ -520,47 +520,47 @@ intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv)
> >   #endif
> >   	/* Get some space for it */
> > -	dev_priv->mch_res.name = "i915 MCHBAR";
> > -	dev_priv->mch_res.flags = IORESOURCE_MEM;
> > -	ret = pci_bus_alloc_resource(dev_priv->bridge_dev->bus,
> > -				     &dev_priv->mch_res,
> > +	i915->mch_res.name = "i915 MCHBAR";
> > +	i915->mch_res.flags = IORESOURCE_MEM;
> > +	ret = pci_bus_alloc_resource(i915->bridge_dev->bus,
> > +				     &i915->mch_res,
> >   				     MCHBAR_SIZE, MCHBAR_SIZE,
> >   				     PCIBIOS_MIN_MEM,
> >   				     0, pcibios_align_resource,
> > -				     dev_priv->bridge_dev);
> > +				     i915->bridge_dev);
> >   	if (ret) {
> >   		DRM_DEBUG_DRIVER("failed bus alloc: %d\n", ret);
> > -		dev_priv->mch_res.start = 0;
> > +		i915->mch_res.start = 0;
> >   		return ret;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > -		pci_write_config_dword(dev_priv->bridge_dev, reg + 4,
> > -				       upper_32_bits(dev_priv->mch_res.start));
> > +	if (INTEL_GEN(i915) >= 4)
> > +		pci_write_config_dword(i915->bridge_dev, reg + 4,
> > +				       upper_32_bits(i915->mch_res.start));
> > -	pci_write_config_dword(dev_priv->bridge_dev, reg,
> > -			       lower_32_bits(dev_priv->mch_res.start));
> > +	pci_write_config_dword(i915->bridge_dev, reg,
> > +			       lower_32_bits(i915->mch_res.start));
> >   	return 0;
> >   }
> >   /* Setup MCHBAR if possible, return true if we should disable it again */
> >   static void
> > -intel_setup_mchbar(struct drm_i915_private *dev_priv)
> > +intel_setup_mchbar(struct drm_i915_private *i915)
> >   {
> > -	int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> > +	int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> >   	u32 temp;
> >   	bool enabled;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		return;
> > -	dev_priv->mchbar_need_disable = false;
> > +	i915->mchbar_need_disable = false;
> > -	if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> > -		pci_read_config_dword(dev_priv->bridge_dev, DEVEN, &temp);
> > +	if (IS_I915G(i915) || IS_I915GM(i915)) {
> > +		pci_read_config_dword(i915->bridge_dev, DEVEN, &temp);
> >   		enabled = !!(temp & DEVEN_MCHBAR_EN);
> >   	} else {
> > -		pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp);
> > +		pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp);
> >   		enabled = temp & 1;
> >   	}
> > @@ -568,56 +568,56 @@ intel_setup_mchbar(struct drm_i915_private *dev_priv)
> >   	if (enabled)
> >   		return;
> > -	if (intel_alloc_mchbar_resource(dev_priv))
> > +	if (intel_alloc_mchbar_resource(i915))
> >   		return;
> > -	dev_priv->mchbar_need_disable = true;
> > +	i915->mchbar_need_disable = true;
> >   	/* Space is allocated or reserved, so enable it. */
> > -	if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> > -		pci_write_config_dword(dev_priv->bridge_dev, DEVEN,
> > +	if (IS_I915G(i915) || IS_I915GM(i915)) {
> > +		pci_write_config_dword(i915->bridge_dev, DEVEN,
> >   				       temp | DEVEN_MCHBAR_EN);
> >   	} else {
> > -		pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp);
> > -		pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg, temp | 1);
> > +		pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp);
> > +		pci_write_config_dword(i915->bridge_dev, mchbar_reg, temp | 1);
> >   	}
> >   }
> >   static void
> > -intel_teardown_mchbar(struct drm_i915_private *dev_priv)
> > +intel_teardown_mchbar(struct drm_i915_private *i915)
> >   {
> > -	int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> > +	int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915;
> > -	if (dev_priv->mchbar_need_disable) {
> > -		if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> > +	if (i915->mchbar_need_disable) {
> > +		if (IS_I915G(i915) || IS_I915GM(i915)) {
> >   			u32 deven_val;
> > -			pci_read_config_dword(dev_priv->bridge_dev, DEVEN,
> > +			pci_read_config_dword(i915->bridge_dev, DEVEN,
> >   					      &deven_val);
> >   			deven_val &= ~DEVEN_MCHBAR_EN;
> > -			pci_write_config_dword(dev_priv->bridge_dev, DEVEN,
> > +			pci_write_config_dword(i915->bridge_dev, DEVEN,
> >   					       deven_val);
> >   		} else {
> >   			u32 mchbar_val;
> > -			pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg,
> > +			pci_read_config_dword(i915->bridge_dev, mchbar_reg,
> >   					      &mchbar_val);
> >   			mchbar_val &= ~1;
> > -			pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg,
> > +			pci_write_config_dword(i915->bridge_dev, mchbar_reg,
> >   					       mchbar_val);
> >   		}
> >   	}
> > -	if (dev_priv->mch_res.start)
> > -		release_resource(&dev_priv->mch_res);
> > +	if (i915->mch_res.start)
> > +		release_resource(&i915->mch_res);
> >   }
> >   /* true = enable decode, false = disable decoder */
> >   static unsigned int i915_vga_set_decode(void *cookie, bool state)
> >   {
> > -	struct drm_i915_private *dev_priv = cookie;
> > +	struct drm_i915_private *i915 = cookie;
> > -	intel_modeset_vga_set_state(dev_priv, state);
> > +	intel_modeset_vga_set_state(i915, state);
> >   	if (state)
> >   		return VGA_RSRC_LEGACY_IO | VGA_RSRC_LEGACY_MEM |
> >   		       VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM;
> > @@ -668,21 +668,21 @@ static const struct vga_switcheroo_client_ops i915_switcheroo_ops = {
> >   static int i915_load_modeset_init(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int ret;
> >   	if (i915_inject_load_failure())
> >   		return -ENODEV;
> > -	if (HAS_DISPLAY(dev_priv)) {
> > -		ret = drm_vblank_init(&dev_priv->drm,
> > -				      INTEL_INFO(dev_priv)->num_pipes);
> > +	if (HAS_DISPLAY(i915)) {
> > +		ret = drm_vblank_init(&i915->drm,
> > +				      INTEL_INFO(i915)->num_pipes);
> >   		if (ret)
> >   			goto out;
> >   	}
> > -	intel_bios_init(dev_priv);
> > +	intel_bios_init(i915);
> >   	/* If we have > 1 VGA cards, then we need to arbitrate access
> >   	 * to the common VGA resources.
> > @@ -691,7 +691,7 @@ static int i915_load_modeset_init(struct drm_device *dev)
> >   	 * then we do not take part in VGA arbitration and the
> >   	 * vga_client_register() fails with -ENODEV.
> >   	 */
> > -	ret = vga_client_register(pdev, dev_priv, NULL, i915_vga_set_decode);
> > +	ret = vga_client_register(pdev, i915, NULL, i915_vga_set_decode);
> >   	if (ret && ret != -ENODEV)
> >   		goto out;
> > @@ -702,17 +702,17 @@ static int i915_load_modeset_init(struct drm_device *dev)
> >   		goto cleanup_vga_client;
> >   	/* must happen before intel_power_domains_init_hw() on VLV/CHV */
> > -	intel_update_rawclk(dev_priv);
> > +	intel_update_rawclk(i915);
> > -	intel_power_domains_init_hw(dev_priv, false);
> > +	intel_power_domains_init_hw(i915, false);
> > -	intel_csr_ucode_init(dev_priv);
> > +	intel_csr_ucode_init(i915);
> > -	ret = intel_irq_install(dev_priv);
> > +	ret = intel_irq_install(i915);
> >   	if (ret)
> >   		goto cleanup_csr;
> > -	intel_gmbus_setup(dev_priv);
> > +	intel_gmbus_setup(i915);
> >   	/* Important: The output setup functions called by modeset_init need
> >   	 * working irqs for e.g. gmbus and dp aux transfers. */
> > @@ -720,13 +720,13 @@ static int i915_load_modeset_init(struct drm_device *dev)
> >   	if (ret)
> >   		goto cleanup_irq;
> > -	ret = i915_gem_init(dev_priv);
> > +	ret = i915_gem_init(i915);
> >   	if (ret)
> >   		goto cleanup_modeset;
> > -	intel_overlay_setup(dev_priv);
> > +	intel_overlay_setup(i915);
> > -	if (!HAS_DISPLAY(dev_priv))
> > +	if (!HAS_DISPLAY(i915))
> >   		return 0;
> >   	ret = intel_fbdev_init(dev);
> > @@ -734,24 +734,24 @@ static int i915_load_modeset_init(struct drm_device *dev)
> >   		goto cleanup_gem;
> >   	/* Only enable hotplug handling once the fbdev is fully set up. */
> > -	intel_hpd_init(dev_priv);
> > +	intel_hpd_init(i915);
> > -	intel_init_ipc(dev_priv);
> > +	intel_init_ipc(i915);
> >   	return 0;
> >   cleanup_gem:
> > -	i915_gem_suspend(dev_priv);
> > -	i915_gem_fini_hw(dev_priv);
> > -	i915_gem_fini(dev_priv);
> > +	i915_gem_suspend(i915);
> > +	i915_gem_fini_hw(i915);
> > +	i915_gem_fini(i915);
> >   cleanup_modeset:
> >   	intel_modeset_cleanup(dev);
> >   cleanup_irq:
> >   	drm_irq_uninstall(dev);
> > -	intel_gmbus_teardown(dev_priv);
> > +	intel_gmbus_teardown(i915);
> >   cleanup_csr:
> > -	intel_csr_ucode_fini(dev_priv);
> > -	intel_power_domains_fini_hw(dev_priv);
> > +	intel_csr_ucode_fini(i915);
> > +	intel_power_domains_fini_hw(i915);
> >   	vga_switcheroo_unregister_client(pdev);
> >   cleanup_vga_client:
> >   	vga_client_register(pdev, NULL, NULL, NULL);
> > @@ -759,11 +759,11 @@ static int i915_load_modeset_init(struct drm_device *dev)
> >   	return ret;
> >   }
> > -static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv)
> > +static int i915_kick_out_firmware_fb(struct drm_i915_private *i915)
> >   {
> >   	struct apertures_struct *ap;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	bool primary;
> >   	int ret;
> > @@ -784,22 +784,22 @@ static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv)
> >   	return ret;
> >   }
> > -static void intel_init_dpio(struct drm_i915_private *dev_priv)
> > +static void intel_init_dpio(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * IOSF_PORT_DPIO is used for VLV x2 PHY (DP/HDMI B and C),
> >   	 * CHV x1 PHY (DP/HDMI D)
> >   	 * IOSF_PORT_DPIO_2 is used for CHV x2 PHY (DP/HDMI B and C)
> >   	 */
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO_2;
> >   		DPIO_PHY_IOSF_PORT(DPIO_PHY1) = IOSF_PORT_DPIO;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO;
> >   	}
> >   }
> > -static int i915_workqueues_init(struct drm_i915_private *dev_priv)
> > +static int i915_workqueues_init(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * The i915 workqueue is primarily used for batched retirement of
> > @@ -815,18 +815,18 @@ static int i915_workqueues_init(struct drm_i915_private *dev_priv)
> >   	 * so there is no point in running more than one instance of the
> >   	 * workqueue at any time.  Use an ordered one.
> >   	 */
> > -	dev_priv->wq = alloc_ordered_workqueue("i915", 0);
> > -	if (dev_priv->wq == NULL)
> > +	i915->wq = alloc_ordered_workqueue("i915", 0);
> > +	if (i915->wq == NULL)
> >   		goto out_err;
> > -	dev_priv->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0);
> > -	if (dev_priv->hotplug.dp_wq == NULL)
> > +	i915->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0);
> > +	if (i915->hotplug.dp_wq == NULL)
> >   		goto out_free_wq;
> >   	return 0;
> >   out_free_wq:
> > -	destroy_workqueue(dev_priv->wq);
> > +	destroy_workqueue(i915->wq);
> >   out_err:
> >   	DRM_ERROR("Failed to allocate workqueues.\n");
> > @@ -842,10 +842,10 @@ static void i915_engines_cleanup(struct drm_i915_private *i915)
> >   		kfree(engine);
> >   }
> > -static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv)
> > +static void i915_workqueues_cleanup(struct drm_i915_private *i915)
> >   {
> > -	destroy_workqueue(dev_priv->hotplug.dp_wq);
> > -	destroy_workqueue(dev_priv->wq);
> > +	destroy_workqueue(i915->hotplug.dp_wq);
> > +	destroy_workqueue(i915->wq);
> >   }
> >   /*
> > @@ -858,14 +858,14 @@ static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv)
> >    * (workarounds have a habit of persisting!). Anything older than that
> >    * should be removed along with the complications they introduce.
> >    */
> > -static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
> > +static void intel_detect_preproduction_hw(struct drm_i915_private *i915)
> >   {
> >   	bool pre = false;
> > -	pre |= IS_HSW_EARLY_SDV(dev_priv);
> > -	pre |= IS_SKL_REVID(dev_priv, 0, SKL_REVID_F0);
> > -	pre |= IS_BXT_REVID(dev_priv, 0, BXT_REVID_B_LAST);
> > -	pre |= IS_KBL_REVID(dev_priv, 0, KBL_REVID_A0);
> > +	pre |= IS_HSW_EARLY_SDV(i915);
> > +	pre |= IS_SKL_REVID(i915, 0, SKL_REVID_F0);
> > +	pre |= IS_BXT_REVID(i915, 0, BXT_REVID_B_LAST);
> > +	pre |= IS_KBL_REVID(i915, 0, KBL_REVID_A0);
> >   	if (pre) {
> >   		DRM_ERROR("This is a pre-production stepping. "
> > @@ -876,7 +876,7 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
> >   /**
> >    * i915_driver_init_early - setup state not requiring device access
> > - * @dev_priv: device private
> > + * @i915: device private
> >    *
> >    * Initialize everything that is a "SW-only" state, that is state not
> >    * requiring accessing the device or exposing the driver via kernel internal
> > @@ -884,152 +884,152 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv)
> >    * system memory allocation, setting up device specific attributes and
> >    * function hooks not requiring accessing the device.
> >    */
> > -static int i915_driver_init_early(struct drm_i915_private *dev_priv)
> > +static int i915_driver_init_early(struct drm_i915_private *i915)
> >   {
> >   	int ret = 0;
> >   	if (i915_inject_load_failure())
> >   		return -ENODEV;
> > -	intel_device_info_subplatform_init(dev_priv);
> > +	intel_device_info_subplatform_init(i915);
> > -	intel_uncore_init_early(&dev_priv->uncore);
> > +	intel_uncore_init_early(&i915->uncore);
> > -	spin_lock_init(&dev_priv->irq_lock);
> > -	spin_lock_init(&dev_priv->gpu_error.lock);
> > -	mutex_init(&dev_priv->backlight_lock);
> > +	spin_lock_init(&i915->irq_lock);
> > +	spin_lock_init(&i915->gpu_error.lock);
> > +	mutex_init(&i915->backlight_lock);
> > -	mutex_init(&dev_priv->sb_lock);
> > -	pm_qos_add_request(&dev_priv->sb_qos,
> > +	mutex_init(&i915->sb_lock);
> > +	pm_qos_add_request(&i915->sb_qos,
> >   			   PM_QOS_CPU_DMA_LATENCY, PM_QOS_DEFAULT_VALUE);
> > -	mutex_init(&dev_priv->av_mutex);
> > -	mutex_init(&dev_priv->wm.wm_mutex);
> > -	mutex_init(&dev_priv->pps_mutex);
> > -	mutex_init(&dev_priv->hdcp_comp_mutex);
> > +	mutex_init(&i915->av_mutex);
> > +	mutex_init(&i915->wm.wm_mutex);
> > +	mutex_init(&i915->pps_mutex);
> > +	mutex_init(&i915->hdcp_comp_mutex);
> > -	i915_memcpy_init_early(dev_priv);
> > -	intel_runtime_pm_init_early(dev_priv);
> > +	i915_memcpy_init_early(i915);
> > +	intel_runtime_pm_init_early(i915);
> > -	ret = i915_workqueues_init(dev_priv);
> > +	ret = i915_workqueues_init(i915);
> >   	if (ret < 0)
> >   		goto err_engines;
> > -	ret = i915_gem_init_early(dev_priv);
> > +	ret = i915_gem_init_early(i915);
> >   	if (ret < 0)
> >   		goto err_workqueues;
> >   	/* This must be called before any calls to HAS_PCH_* */
> > -	intel_detect_pch(dev_priv);
> > +	intel_detect_pch(i915);
> > -	intel_wopcm_init_early(&dev_priv->wopcm);
> > -	intel_uc_init_early(dev_priv);
> > -	intel_pm_setup(dev_priv);
> > -	intel_init_dpio(dev_priv);
> > -	ret = intel_power_domains_init(dev_priv);
> > +	intel_wopcm_init_early(&i915->wopcm);
> > +	intel_uc_init_early(i915);
> > +	intel_pm_setup(i915);
> > +	intel_init_dpio(i915);
> > +	ret = intel_power_domains_init(i915);
> >   	if (ret < 0)
> >   		goto err_uc;
> > -	intel_irq_init(dev_priv);
> > -	intel_hangcheck_init(dev_priv);
> > -	intel_init_display_hooks(dev_priv);
> > -	intel_init_clock_gating_hooks(dev_priv);
> > -	intel_init_audio_hooks(dev_priv);
> > -	intel_display_crc_init(dev_priv);
> > +	intel_irq_init(i915);
> > +	intel_hangcheck_init(i915);
> > +	intel_init_display_hooks(i915);
> > +	intel_init_clock_gating_hooks(i915);
> > +	intel_init_audio_hooks(i915);
> > +	intel_display_crc_init(i915);
> > -	intel_detect_preproduction_hw(dev_priv);
> > +	intel_detect_preproduction_hw(i915);
> >   	return 0;
> >   err_uc:
> > -	intel_uc_cleanup_early(dev_priv);
> > -	i915_gem_cleanup_early(dev_priv);
> > +	intel_uc_cleanup_early(i915);
> > +	i915_gem_cleanup_early(i915);
> >   err_workqueues:
> > -	i915_workqueues_cleanup(dev_priv);
> > +	i915_workqueues_cleanup(i915);
> >   err_engines:
> > -	i915_engines_cleanup(dev_priv);
> > +	i915_engines_cleanup(i915);
> >   	return ret;
> >   }
> >   /**
> >    * i915_driver_cleanup_early - cleanup the setup done in i915_driver_init_early()
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -static void i915_driver_cleanup_early(struct drm_i915_private *dev_priv)
> > +static void i915_driver_cleanup_early(struct drm_i915_private *i915)
> >   {
> > -	intel_irq_fini(dev_priv);
> > -	intel_power_domains_cleanup(dev_priv);
> > -	intel_uc_cleanup_early(dev_priv);
> > -	i915_gem_cleanup_early(dev_priv);
> > -	i915_workqueues_cleanup(dev_priv);
> > -	i915_engines_cleanup(dev_priv);
> > +	intel_irq_fini(i915);
> > +	intel_power_domains_cleanup(i915);
> > +	intel_uc_cleanup_early(i915);
> > +	i915_gem_cleanup_early(i915);
> > +	i915_workqueues_cleanup(i915);
> > +	i915_engines_cleanup(i915);
> > -	pm_qos_remove_request(&dev_priv->sb_qos);
> > -	mutex_destroy(&dev_priv->sb_lock);
> > +	pm_qos_remove_request(&i915->sb_qos);
> > +	mutex_destroy(&i915->sb_lock);
> >   }
> >   /**
> >    * i915_driver_init_mmio - setup device MMIO
> > - * @dev_priv: device private
> > + * @i915: device private
> >    *
> >    * Setup minimal device state necessary for MMIO accesses later in the
> >    * initialization sequence. The setup here should avoid any other device-wide
> >    * side effects or exposing the driver via kernel internal or user space
> >    * interfaces.
> >    */
> > -static int i915_driver_init_mmio(struct drm_i915_private *dev_priv)
> > +static int i915_driver_init_mmio(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> >   	if (i915_inject_load_failure())
> >   		return -ENODEV;
> > -	if (i915_get_bridge_dev(dev_priv))
> > +	if (i915_get_bridge_dev(i915))
> >   		return -EIO;
> > -	ret = intel_uncore_init_mmio(&dev_priv->uncore);
> > +	ret = intel_uncore_init_mmio(&i915->uncore);
> >   	if (ret < 0)
> >   		goto err_bridge;
> >   	/* Try to make sure MCHBAR is enabled before poking at it */
> > -	intel_setup_mchbar(dev_priv);
> > +	intel_setup_mchbar(i915);
> > -	intel_device_info_init_mmio(dev_priv);
> > +	intel_device_info_init_mmio(i915);
> > -	intel_uncore_prune_mmio_domains(&dev_priv->uncore);
> > +	intel_uncore_prune_mmio_domains(&i915->uncore);
> > -	intel_uc_init_mmio(dev_priv);
> > +	intel_uc_init_mmio(i915);
> > -	ret = intel_engines_init_mmio(dev_priv);
> > +	ret = intel_engines_init_mmio(i915);
> >   	if (ret)
> >   		goto err_uncore;
> > -	i915_gem_init_mmio(dev_priv);
> > +	i915_gem_init_mmio(i915);
> >   	return 0;
> >   err_uncore:
> > -	intel_teardown_mchbar(dev_priv);
> > -	intel_uncore_fini_mmio(&dev_priv->uncore);
> > +	intel_teardown_mchbar(i915);
> > +	intel_uncore_fini_mmio(&i915->uncore);
> >   err_bridge:
> > -	pci_dev_put(dev_priv->bridge_dev);
> > +	pci_dev_put(i915->bridge_dev);
> >   	return ret;
> >   }
> >   /**
> >    * i915_driver_cleanup_mmio - cleanup the setup done in i915_driver_init_mmio()
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -static void i915_driver_cleanup_mmio(struct drm_i915_private *dev_priv)
> > +static void i915_driver_cleanup_mmio(struct drm_i915_private *i915)
> >   {
> > -	intel_teardown_mchbar(dev_priv);
> > -	intel_uncore_fini_mmio(&dev_priv->uncore);
> > -	pci_dev_put(dev_priv->bridge_dev);
> > +	intel_teardown_mchbar(i915);
> > +	intel_uncore_fini_mmio(&i915->uncore);
> > +	pci_dev_put(i915->bridge_dev);
> >   }
> > -static void intel_sanitize_options(struct drm_i915_private *dev_priv)
> > +static void intel_sanitize_options(struct drm_i915_private *i915)
> >   {
> > -	intel_gvt_sanitize_options(dev_priv);
> > +	intel_gvt_sanitize_options(i915);
> >   }
> >   #define DRAM_TYPE_STR(type) [INTEL_DRAM_ ## type] = #type
> > @@ -1131,11 +1131,11 @@ skl_is_16gb_dimm(const struct dram_dimm_info *dimm)
> >   }
> >   static void
> > -skl_dram_get_dimm_info(struct drm_i915_private *dev_priv,
> > +skl_dram_get_dimm_info(struct drm_i915_private *i915,
> >   		       struct dram_dimm_info *dimm,
> >   		       int channel, char dimm_name, u16 val)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 10) {
> > +	if (INTEL_GEN(i915) >= 10) {
> >   		dimm->size = cnl_get_dimm_size(val);
> >   		dimm->width = cnl_get_dimm_width(val);
> >   		dimm->ranks = cnl_get_dimm_ranks(val);
> > @@ -1151,13 +1151,13 @@ skl_dram_get_dimm_info(struct drm_i915_private *dev_priv,
> >   }
> >   static int
> > -skl_dram_get_channel_info(struct drm_i915_private *dev_priv,
> > +skl_dram_get_channel_info(struct drm_i915_private *i915,
> >   			  struct dram_channel_info *ch,
> >   			  int channel, u32 val)
> >   {
> > -	skl_dram_get_dimm_info(dev_priv, &ch->dimm_l,
> > +	skl_dram_get_dimm_info(i915, &ch->dimm_l,
> >   			       channel, 'L', val & 0xffff);
> > -	skl_dram_get_dimm_info(dev_priv, &ch->dimm_s,
> > +	skl_dram_get_dimm_info(i915, &ch->dimm_s,
> >   			       channel, 'S', val >> 16);
> >   	if (ch->dimm_l.size == 0 && ch->dimm_s.size == 0) {
> > @@ -1192,20 +1192,20 @@ intel_is_dram_symmetric(const struct dram_channel_info *ch0,
> >   }
> >   static int
> > -skl_dram_get_channels_info(struct drm_i915_private *dev_priv)
> > +skl_dram_get_channels_info(struct drm_i915_private *i915)
> >   {
> > -	struct dram_info *dram_info = &dev_priv->dram_info;
> > +	struct dram_info *dram_info = &i915->dram_info;
> >   	struct dram_channel_info ch0 = {}, ch1 = {};
> >   	u32 val;
> >   	int ret;
> >   	val = I915_READ(SKL_MAD_DIMM_CH0_0_0_0_MCHBAR_MCMAIN);
> > -	ret = skl_dram_get_channel_info(dev_priv, &ch0, 0, val);
> > +	ret = skl_dram_get_channel_info(i915, &ch0, 0, val);
> >   	if (ret == 0)
> >   		dram_info->num_channels++;
> >   	val = I915_READ(SKL_MAD_DIMM_CH1_0_0_0_MCHBAR_MCMAIN);
> > -	ret = skl_dram_get_channel_info(dev_priv, &ch1, 1, val);
> > +	ret = skl_dram_get_channel_info(i915, &ch1, 1, val);
> >   	if (ret == 0)
> >   		dram_info->num_channels++;
> > @@ -1239,7 +1239,7 @@ skl_dram_get_channels_info(struct drm_i915_private *dev_priv)
> >   }
> >   static enum intel_dram_type
> > -skl_get_dram_type(struct drm_i915_private *dev_priv)
> > +skl_get_dram_type(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -1261,16 +1261,16 @@ skl_get_dram_type(struct drm_i915_private *dev_priv)
> >   }
> >   static int
> > -skl_get_dram_info(struct drm_i915_private *dev_priv)
> > +skl_get_dram_info(struct drm_i915_private *i915)
> >   {
> > -	struct dram_info *dram_info = &dev_priv->dram_info;
> > +	struct dram_info *dram_info = &i915->dram_info;
> >   	u32 mem_freq_khz, val;
> >   	int ret;
> > -	dram_info->type = skl_get_dram_type(dev_priv);
> > +	dram_info->type = skl_get_dram_type(i915);
> >   	DRM_DEBUG_KMS("DRAM type: %s\n", intel_dram_type_str(dram_info->type));
> > -	ret = skl_dram_get_channels_info(dev_priv);
> > +	ret = skl_dram_get_channels_info(i915);
> >   	if (ret)
> >   		return ret;
> > @@ -1370,9 +1370,9 @@ static void bxt_get_dimm_info(struct dram_dimm_info *dimm,
> >   }
> >   static int
> > -bxt_get_dram_info(struct drm_i915_private *dev_priv)
> > +bxt_get_dram_info(struct drm_i915_private *i915)
> >   {
> > -	struct dram_info *dram_info = &dev_priv->dram_info;
> > +	struct dram_info *dram_info = &i915->dram_info;
> >   	u32 dram_channels;
> >   	u32 mem_freq_khz, val;
> >   	u8 num_active_channels;
> > @@ -1443,9 +1443,9 @@ bxt_get_dram_info(struct drm_i915_private *dev_priv)
> >   }
> >   static void
> > -intel_get_dram_info(struct drm_i915_private *dev_priv)
> > +intel_get_dram_info(struct drm_i915_private *i915)
> >   {
> > -	struct dram_info *dram_info = &dev_priv->dram_info;
> > +	struct dram_info *dram_info = &i915->dram_info;
> >   	int ret;
> >   	/*
> > @@ -1453,15 +1453,15 @@ intel_get_dram_info(struct drm_i915_private *dev_priv)
> >   	 * This is only used for the level 0 watermark latency
> >   	 * w/a which does not apply to bxt/glk.
> >   	 */
> > -	dram_info->is_16gb_dimm = !IS_GEN9_LP(dev_priv);
> > +	dram_info->is_16gb_dimm = !IS_GEN9_LP(i915);
> > -	if (INTEL_GEN(dev_priv) < 9)
> > +	if (INTEL_GEN(i915) < 9)
> >   		return;
> > -	if (IS_GEN9_LP(dev_priv))
> > -		ret = bxt_get_dram_info(dev_priv);
> > +	if (IS_GEN9_LP(i915))
> > +		ret = bxt_get_dram_info(i915);
> >   	else
> > -		ret = skl_get_dram_info(dev_priv);
> > +		ret = skl_get_dram_info(i915);
> >   	if (ret)
> >   		return;
> > @@ -1473,7 +1473,7 @@ intel_get_dram_info(struct drm_i915_private *dev_priv)
> >   		      dram_info->ranks, yesno(dram_info->is_16gb_dimm));
> >   }
> > -static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap)
> > +static u32 gen9_edram_size_mb(struct drm_i915_private *i915, u32 cap)
> >   {
> >   	const unsigned int ways[8] = { 4, 8, 12, 16, 16, 16, 16, 16 };
> >   	const unsigned int sets[4] = { 1, 1, 2, 2 };
> > @@ -1483,16 +1483,16 @@ static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap)
> >   		sets[EDRAM_SETS_IDX(cap)];
> >   }
> > -static void edram_detect(struct drm_i915_private *dev_priv)
> > +static void edram_detect(struct drm_i915_private *i915)
> >   {
> >   	u32 edram_cap = 0;
> > -	if (!(IS_HASWELL(dev_priv) ||
> > -	      IS_BROADWELL(dev_priv) ||
> > -	      INTEL_GEN(dev_priv) >= 9))
> > +	if (!(IS_HASWELL(i915) ||
> > +	      IS_BROADWELL(i915) ||
> > +	      INTEL_GEN(i915) >= 9))
> >   		return;
> > -	edram_cap = __raw_uncore_read32(&dev_priv->uncore, HSW_EDRAM_CAP);
> > +	edram_cap = __raw_uncore_read32(&i915->uncore, HSW_EDRAM_CAP);
> >   	/* NB: We can't write IDICR yet because we don't have gt funcs set up */
> > @@ -1503,63 +1503,63 @@ static void edram_detect(struct drm_i915_private *dev_priv)
> >   	 * The needed capability bits for size calculation are not there with
> >   	 * pre gen9 so return 128MB always.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 9)
> > -		dev_priv->edram_size_mb = 128;
> > +	if (INTEL_GEN(i915) < 9)
> > +		i915->edram_size_mb = 128;
> >   	else
> > -		dev_priv->edram_size_mb =
> > -			gen9_edram_size_mb(dev_priv, edram_cap);
> > +		i915->edram_size_mb =
> > +			gen9_edram_size_mb(i915, edram_cap);
> > -	DRM_INFO("Found %uMB of eDRAM\n", dev_priv->edram_size_mb);
> > +	DRM_INFO("Found %uMB of eDRAM\n", i915->edram_size_mb);
> >   }
> >   /**
> >    * i915_driver_init_hw - setup state requiring device access
> > - * @dev_priv: device private
> > + * @i915: device private
> >    *
> >    * Setup state that requires accessing the device, but doesn't require
> >    * exposing the driver via kernel internal or userspace interfaces.
> >    */
> > -static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> > +static int i915_driver_init_hw(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int ret;
> >   	if (i915_inject_load_failure())
> >   		return -ENODEV;
> > -	intel_device_info_runtime_init(dev_priv);
> > +	intel_device_info_runtime_init(i915);
> > -	if (HAS_PPGTT(dev_priv)) {
> > -		if (intel_vgpu_active(dev_priv) &&
> > -		    !intel_vgpu_has_full_ppgtt(dev_priv)) {
> > -			i915_report_error(dev_priv,
> > +	if (HAS_PPGTT(i915)) {
> > +		if (intel_vgpu_active(i915) &&
> > +		    !intel_vgpu_has_full_ppgtt(i915)) {
> > +			i915_report_error(i915,
> >   					  "incompatible vGPU found, support for isolated ppGTT required\n");
> >   			return -ENXIO;
> >   		}
> >   	}
> > -	if (HAS_EXECLISTS(dev_priv)) {
> > +	if (HAS_EXECLISTS(i915)) {
> >   		/*
> >   		 * Older GVT emulation depends upon intercepting CSB mmio,
> >   		 * which we no longer use, preferring to use the HWSP cache
> >   		 * instead.
> >   		 */
> > -		if (intel_vgpu_active(dev_priv) &&
> > -		    !intel_vgpu_has_hwsp_emulation(dev_priv)) {
> > -			i915_report_error(dev_priv,
> > +		if (intel_vgpu_active(i915) &&
> > +		    !intel_vgpu_has_hwsp_emulation(i915)) {
> > +			i915_report_error(i915,
> >   					  "old vGPU host found, support for HWSP emulation required\n");
> >   			return -ENXIO;
> >   		}
> >   	}
> > -	intel_sanitize_options(dev_priv);
> > +	intel_sanitize_options(i915);
> >   	/* needs to be done before ggtt probe */
> > -	edram_detect(dev_priv);
> > +	edram_detect(i915);
> > -	i915_perf_init(dev_priv);
> > +	i915_perf_init(i915);
> > -	ret = i915_ggtt_probe_hw(dev_priv);
> > +	ret = i915_ggtt_probe_hw(i915);
> >   	if (ret)
> >   		goto err_perf;
> > @@ -1567,7 +1567,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   	 * WARNING: Apparently we must kick fbdev drivers before vgacon,
> >   	 * otherwise the vga fbdev driver falls over.
> >   	 */
> > -	ret = i915_kick_out_firmware_fb(dev_priv);
> > +	ret = i915_kick_out_firmware_fb(i915);
> >   	if (ret) {
> >   		DRM_ERROR("failed to remove conflicting framebuffer drivers\n");
> >   		goto err_ggtt;
> > @@ -1579,11 +1579,11 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   		goto err_ggtt;
> >   	}
> > -	ret = i915_ggtt_init_hw(dev_priv);
> > +	ret = i915_ggtt_init_hw(i915);
> >   	if (ret)
> >   		goto err_ggtt;
> > -	ret = i915_ggtt_enable_hw(dev_priv);
> > +	ret = i915_ggtt_enable_hw(i915);
> >   	if (ret) {
> >   		DRM_ERROR("failed to enable GGTT\n");
> >   		goto err_ggtt;
> > @@ -1592,7 +1592,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   	pci_set_master(pdev);
> >   	/* overlay on gen2 is broken and can't address above 1G */
> > -	if (IS_GEN(dev_priv, 2)) {
> > +	if (IS_GEN(i915, 2)) {
> >   		ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(30));
> >   		if (ret) {
> >   			DRM_ERROR("failed to set DMA mask\n");
> > @@ -1609,7 +1609,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   	 * behaviour if any general state is accessed within a page above 4GB,
> >   	 * which also needs to be handled carefully.
> >   	 */
> > -	if (IS_I965G(dev_priv) || IS_I965GM(dev_priv)) {
> > +	if (IS_I965G(i915) || IS_I965GM(i915)) {
> >   		ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
> >   		if (ret) {
> > @@ -1619,13 +1619,13 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   		}
> >   	}
> > -	pm_qos_add_request(&dev_priv->pm_qos, PM_QOS_CPU_DMA_LATENCY,
> > +	pm_qos_add_request(&i915->pm_qos, PM_QOS_CPU_DMA_LATENCY,
> >   			   PM_QOS_DEFAULT_VALUE);
> > -	intel_uncore_sanitize(dev_priv);
> > +	intel_uncore_sanitize(i915);
> > -	intel_gt_init_workarounds(dev_priv);
> > -	i915_gem_load_init_fences(dev_priv);
> > +	intel_gt_init_workarounds(i915);
> > +	i915_gem_load_init_fences(i915);
> >   	/* On the 945G/GM, the chipset reports the MSI capability on the
> >   	 * integrated graphics even though the support isn't actually there
> > @@ -1646,94 +1646,94 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv)
> >   	 * device. The kernel then disables that interrupt source and so
> >   	 * prevents the other device from working properly.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 5) {
> > +	if (INTEL_GEN(i915) >= 5) {
> >   		if (pci_enable_msi(pdev) < 0)
> >   			DRM_DEBUG_DRIVER("can't enable MSI");
> >   	}
> > -	ret = intel_gvt_init(dev_priv);
> > +	ret = intel_gvt_init(i915);
> >   	if (ret)
> >   		goto err_msi;
> > -	intel_opregion_setup(dev_priv);
> > +	intel_opregion_setup(i915);
> >   	/*
> >   	 * Fill the dram structure to get the system raw bandwidth and
> >   	 * dram info. This will be used for memory latency calculation.
> >   	 */
> > -	intel_get_dram_info(dev_priv);
> > +	intel_get_dram_info(i915);
> > -	intel_bw_init_hw(dev_priv);
> > +	intel_bw_init_hw(i915);
> >   	return 0;
> >   err_msi:
> >   	if (pdev->msi_enabled)
> >   		pci_disable_msi(pdev);
> > -	pm_qos_remove_request(&dev_priv->pm_qos);
> > +	pm_qos_remove_request(&i915->pm_qos);
> >   err_ggtt:
> > -	i915_ggtt_cleanup_hw(dev_priv);
> > +	i915_ggtt_cleanup_hw(i915);
> >   err_perf:
> > -	i915_perf_fini(dev_priv);
> > +	i915_perf_fini(i915);
> >   	return ret;
> >   }
> >   /**
> >    * i915_driver_cleanup_hw - cleanup the setup done in i915_driver_init_hw()
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -static void i915_driver_cleanup_hw(struct drm_i915_private *dev_priv)
> > +static void i915_driver_cleanup_hw(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > -	i915_perf_fini(dev_priv);
> > +	i915_perf_fini(i915);
> >   	if (pdev->msi_enabled)
> >   		pci_disable_msi(pdev);
> > -	pm_qos_remove_request(&dev_priv->pm_qos);
> > +	pm_qos_remove_request(&i915->pm_qos);
> >   }
> >   /**
> >    * i915_driver_register - register the driver with the rest of the system
> > - * @dev_priv: device private
> > + * @i915: device private
> >    *
> >    * Perform any steps necessary to make the driver available via kernel
> >    * internal or userspace interfaces.
> >    */
> > -static void i915_driver_register(struct drm_i915_private *dev_priv)
> > +static void i915_driver_register(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> > -	i915_gem_shrinker_register(dev_priv);
> > -	i915_pmu_register(dev_priv);
> > +	i915_gem_shrinker_register(i915);
> > +	i915_pmu_register(i915);
> >   	/*
> >   	 * Notify a valid surface after modesetting,
> >   	 * when running inside a VM.
> >   	 */
> > -	if (intel_vgpu_active(dev_priv))
> > +	if (intel_vgpu_active(i915))
> >   		I915_WRITE(vgtif_reg(display_ready), VGT_DRV_DISPLAY_READY);
> >   	/* Reveal our presence to userspace */
> >   	if (drm_dev_register(dev, 0) == 0) {
> > -		i915_debugfs_register(dev_priv);
> > -		i915_setup_sysfs(dev_priv);
> > +		i915_debugfs_register(i915);
> > +		i915_setup_sysfs(i915);
> >   		/* Depends on sysfs having been initialized */
> > -		i915_perf_register(dev_priv);
> > +		i915_perf_register(i915);
> >   	} else
> >   		DRM_ERROR("Failed to register driver for userspace access!\n");
> > -	if (HAS_DISPLAY(dev_priv)) {
> > +	if (HAS_DISPLAY(i915)) {
> >   		/* Must be done after probing outputs */
> > -		intel_opregion_register(dev_priv);
> > +		intel_opregion_register(i915);
> >   		acpi_video_register();
> >   	}
> > -	if (IS_GEN(dev_priv, 5))
> > -		intel_gpu_ips_init(dev_priv);
> > +	if (IS_GEN(i915, 5))
> > +		intel_gpu_ips_init(i915);
> > -	intel_audio_init(dev_priv);
> > +	intel_audio_init(i915);
> >   	/*
> >   	 * Some ports require correctly set-up hpd registers for detection to
> > @@ -1748,60 +1748,60 @@ static void i915_driver_register(struct drm_i915_private *dev_priv)
> >   	 * We need to coordinate the hotplugs with the asynchronous fbdev
> >   	 * configuration, for which we use the fbdev->async_cookie.
> >   	 */
> > -	if (HAS_DISPLAY(dev_priv))
> > +	if (HAS_DISPLAY(i915))
> >   		drm_kms_helper_poll_init(dev);
> > -	intel_power_domains_enable(dev_priv);
> > -	intel_runtime_pm_enable(dev_priv);
> > +	intel_power_domains_enable(i915);
> > +	intel_runtime_pm_enable(i915);
> >   }
> >   /**
> >    * i915_driver_unregister - cleanup the registration done in i915_driver_regiser()
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -static void i915_driver_unregister(struct drm_i915_private *dev_priv)
> > +static void i915_driver_unregister(struct drm_i915_private *i915)
> >   {
> > -	intel_runtime_pm_disable(dev_priv);
> > -	intel_power_domains_disable(dev_priv);
> > +	intel_runtime_pm_disable(i915);
> > +	intel_power_domains_disable(i915);
> > -	intel_fbdev_unregister(dev_priv);
> > -	intel_audio_deinit(dev_priv);
> > +	intel_fbdev_unregister(i915);
> > +	intel_audio_deinit(i915);
> >   	/*
> >   	 * After flushing the fbdev (incl. a late async config which will
> >   	 * have delayed queuing of a hotplug event), then flush the hotplug
> >   	 * events.
> >   	 */
> > -	drm_kms_helper_poll_fini(&dev_priv->drm);
> > +	drm_kms_helper_poll_fini(&i915->drm);
> >   	intel_gpu_ips_teardown();
> >   	acpi_video_unregister();
> > -	intel_opregion_unregister(dev_priv);
> > +	intel_opregion_unregister(i915);
> > -	i915_perf_unregister(dev_priv);
> > -	i915_pmu_unregister(dev_priv);
> > +	i915_perf_unregister(i915);
> > +	i915_pmu_unregister(i915);
> > -	i915_teardown_sysfs(dev_priv);
> > -	drm_dev_unplug(&dev_priv->drm);
> > +	i915_teardown_sysfs(i915);
> > +	drm_dev_unplug(&i915->drm);
> > -	i915_gem_shrinker_unregister(dev_priv);
> > +	i915_gem_shrinker_unregister(i915);
> >   }
> > -static void i915_welcome_messages(struct drm_i915_private *dev_priv)
> > +static void i915_welcome_messages(struct drm_i915_private *i915)
> >   {
> >   	if (drm_debug & DRM_UT_DRIVER) {
> >   		struct drm_printer p = drm_debug_printer("i915 device info:");
> >   		drm_printf(&p, "pciid=0x%04x rev=0x%02x platform=%s (subplatform=0x%x) gen=%i\n",
> > -			   INTEL_DEVID(dev_priv),
> > -			   INTEL_REVID(dev_priv),
> > -			   intel_platform_name(INTEL_INFO(dev_priv)->platform),
> > -			   intel_subplatform(RUNTIME_INFO(dev_priv),
> > -					     INTEL_INFO(dev_priv)->platform),
> > -			   INTEL_GEN(dev_priv));
> > -
> > -		intel_device_info_dump_flags(INTEL_INFO(dev_priv), &p);
> > -		intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p);
> > +			   INTEL_DEVID(i915),
> > +			   INTEL_REVID(i915),
> > +			   intel_platform_name(INTEL_INFO(i915)->platform),
> > +			   intel_subplatform(RUNTIME_INFO(i915),
> > +					     INTEL_INFO(i915)->platform),
> > +			   INTEL_GEN(i915));
> > +
> > +		intel_device_info_dump_flags(INTEL_INFO(i915), &p);
> > +		intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p);
> >   	}
> >   	if (IS_ENABLED(CONFIG_DRM_I915_DEBUG))
> > @@ -1871,126 +1871,126 @@ int i915_driver_load(struct pci_dev *pdev, const struct pci_device_id *ent)
> >   {
> >   	const struct intel_device_info *match_info =
> >   		(struct intel_device_info *)ent->driver_data;
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	int ret;
> > -	dev_priv = i915_driver_create(pdev, ent);
> > -	if (IS_ERR(dev_priv))
> > -		return PTR_ERR(dev_priv);
> > +	i915 = i915_driver_create(pdev, ent);
> > +	if (IS_ERR(i915))
> > +		return PTR_ERR(i915);
> >   	/* Disable nuclear pageflip by default on pre-ILK */
> >   	if (!i915_modparams.nuclear_pageflip && match_info->gen < 5)
> > -		dev_priv->drm.driver_features &= ~DRIVER_ATOMIC;
> > +		i915->drm.driver_features &= ~DRIVER_ATOMIC;
> >   	ret = pci_enable_device(pdev);
> >   	if (ret)
> >   		goto out_fini;
> > -	ret = i915_driver_init_early(dev_priv);
> > +	ret = i915_driver_init_early(i915);
> >   	if (ret < 0)
> >   		goto out_pci_disable;
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > -	ret = i915_driver_init_mmio(dev_priv);
> > +	ret = i915_driver_init_mmio(i915);
> >   	if (ret < 0)
> >   		goto out_runtime_pm_put;
> > -	ret = i915_driver_init_hw(dev_priv);
> > +	ret = i915_driver_init_hw(i915);
> >   	if (ret < 0)
> >   		goto out_cleanup_mmio;
> > -	ret = i915_load_modeset_init(&dev_priv->drm);
> > +	ret = i915_load_modeset_init(&i915->drm);
> >   	if (ret < 0)
> >   		goto out_cleanup_hw;
> > -	i915_driver_register(dev_priv);
> > +	i915_driver_register(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> > -	i915_welcome_messages(dev_priv);
> > +	i915_welcome_messages(i915);
> >   	return 0;
> >   out_cleanup_hw:
> > -	i915_driver_cleanup_hw(dev_priv);
> > -	i915_ggtt_cleanup_hw(dev_priv);
> > +	i915_driver_cleanup_hw(i915);
> > +	i915_ggtt_cleanup_hw(i915);
> >   out_cleanup_mmio:
> > -	i915_driver_cleanup_mmio(dev_priv);
> > +	i915_driver_cleanup_mmio(i915);
> >   out_runtime_pm_put:
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > -	i915_driver_cleanup_early(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> > +	i915_driver_cleanup_early(i915);
> >   out_pci_disable:
> >   	pci_disable_device(pdev);
> >   out_fini:
> > -	i915_load_error(dev_priv, "Device initialization failed (%d)\n", ret);
> > -	i915_driver_destroy(dev_priv);
> > +	i915_load_error(i915, "Device initialization failed (%d)\n", ret);
> > +	i915_driver_destroy(i915);
> >   	return ret;
> >   }
> >   void i915_driver_unload(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > -	i915_driver_unregister(dev_priv);
> > +	i915_driver_unregister(i915);
> >   	/*
> >   	 * After unregistering the device to prevent any new users, cancel
> >   	 * all in-flight requests so that we can quickly unbind the active
> >   	 * resources.
> >   	 */
> > -	i915_gem_set_wedged(dev_priv);
> > +	i915_gem_set_wedged(i915);
> >   	/* Flush any external code that still may be under the RCU lock */
> >   	synchronize_rcu();
> > -	i915_gem_suspend(dev_priv);
> > +	i915_gem_suspend(i915);
> >   	drm_atomic_helper_shutdown(dev);
> > -	intel_gvt_cleanup(dev_priv);
> > +	intel_gvt_cleanup(i915);
> >   	intel_modeset_cleanup(dev);
> > -	intel_bios_cleanup(dev_priv);
> > +	intel_bios_cleanup(i915);
> >   	vga_switcheroo_unregister_client(pdev);
> >   	vga_client_register(pdev, NULL, NULL, NULL);
> > -	intel_csr_ucode_fini(dev_priv);
> > +	intel_csr_ucode_fini(i915);
> >   	/* Free error state after interrupts are fully disabled. */
> > -	cancel_delayed_work_sync(&dev_priv->gpu_error.hangcheck_work);
> > -	i915_reset_error_state(dev_priv);
> > +	cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
> > +	i915_reset_error_state(i915);
> > -	i915_gem_fini_hw(dev_priv);
> > +	i915_gem_fini_hw(i915);
> > -	intel_power_domains_fini_hw(dev_priv);
> > +	intel_power_domains_fini_hw(i915);
> > -	i915_driver_cleanup_hw(dev_priv);
> > +	i915_driver_cleanup_hw(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   }
> >   static void i915_driver_release(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > -	i915_gem_fini(dev_priv);
> > +	i915_gem_fini(i915);
> > -	i915_ggtt_cleanup_hw(dev_priv);
> > -	i915_driver_cleanup_mmio(dev_priv);
> > +	i915_ggtt_cleanup_hw(i915);
> > +	i915_driver_cleanup_mmio(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > -	intel_runtime_pm_cleanup(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> > +	intel_runtime_pm_cleanup(i915);
> > -	i915_driver_cleanup_early(dev_priv);
> > -	i915_driver_destroy(dev_priv);
> > +	i915_driver_cleanup_early(i915);
> > +	i915_driver_destroy(i915);
> >   }
> >   static int i915_driver_open(struct drm_device *dev, struct drm_file *file)
> > @@ -2035,9 +2035,9 @@ static void i915_driver_postclose(struct drm_device *dev, struct drm_file *file)
> >   	kfree(file_priv);
> >   }
> > -static void intel_suspend_encoders(struct drm_i915_private *dev_priv)
> > +static void intel_suspend_encoders(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_encoder *encoder;
> >   	drm_modeset_lock_all(dev);
> > @@ -2047,11 +2047,11 @@ static void intel_suspend_encoders(struct drm_i915_private *dev_priv)
> >   	drm_modeset_unlock_all(dev);
> >   }
> > -static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> > +static int vlv_resume_prepare(struct drm_i915_private *i915,
> >   			      bool rpm_resume);
> > -static int vlv_suspend_complete(struct drm_i915_private *dev_priv);
> > +static int vlv_suspend_complete(struct drm_i915_private *i915);
> > -static bool suspend_to_idle(struct drm_i915_private *dev_priv)
> > +static bool suspend_to_idle(struct drm_i915_private *i915)
> >   {
> >   #if IS_ENABLED(CONFIG_ACPI_SLEEP)
> >   	if (acpi_target_system_state() < ACPI_STATE_S3)
> > @@ -2077,15 +2077,15 @@ static int i915_drm_prepare(struct drm_device *dev)
> >   static int i915_drm_suspend(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	pci_power_t opregion_target_state;
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	/* We do a lot of poking in a lot of registers, make sure they work
> >   	 * properly. */
> > -	intel_power_domains_disable(dev_priv);
> > +	intel_power_domains_disable(i915);
> >   	drm_kms_helper_poll_disable(dev);
> > @@ -2093,40 +2093,40 @@ static int i915_drm_suspend(struct drm_device *dev)
> >   	intel_display_suspend(dev);
> > -	intel_dp_mst_suspend(dev_priv);
> > +	intel_dp_mst_suspend(i915);
> > -	intel_runtime_pm_disable_interrupts(dev_priv);
> > -	intel_hpd_cancel_work(dev_priv);
> > +	intel_runtime_pm_disable_interrupts(i915);
> > +	intel_hpd_cancel_work(i915);
> > -	intel_suspend_encoders(dev_priv);
> > +	intel_suspend_encoders(i915);
> > -	intel_suspend_hw(dev_priv);
> > +	intel_suspend_hw(i915);
> > -	i915_gem_suspend_gtt_mappings(dev_priv);
> > +	i915_gem_suspend_gtt_mappings(i915);
> > -	i915_save_state(dev_priv);
> > +	i915_save_state(i915);
> > -	opregion_target_state = suspend_to_idle(dev_priv) ? PCI_D1 : PCI_D3cold;
> > -	intel_opregion_suspend(dev_priv, opregion_target_state);
> > +	opregion_target_state = suspend_to_idle(i915) ? PCI_D1 : PCI_D3cold;
> > +	intel_opregion_suspend(i915, opregion_target_state);
> >   	intel_fbdev_set_suspend(dev, FBINFO_STATE_SUSPENDED, true);
> > -	dev_priv->suspend_count++;
> > +	i915->suspend_count++;
> > -	intel_csr_ucode_suspend(dev_priv);
> > +	intel_csr_ucode_suspend(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return 0;
> >   }
> >   static enum i915_drm_suspend_mode
> > -get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate)
> > +get_suspend_mode(struct drm_i915_private *i915, bool hibernate)
> >   {
> >   	if (hibernate)
> >   		return I915_DRM_SUSPEND_HIBERNATE;
> > -	if (suspend_to_idle(dev_priv))
> > +	if (suspend_to_idle(i915))
> >   		return I915_DRM_SUSPEND_IDLE;
> >   	return I915_DRM_SUSPEND_MEM;
> > @@ -2134,30 +2134,30 @@ get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate)
> >   static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int ret;
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > -	i915_gem_suspend_late(dev_priv);
> > +	i915_gem_suspend_late(i915);
> > -	intel_uncore_suspend(&dev_priv->uncore);
> > +	intel_uncore_suspend(&i915->uncore);
> > -	intel_power_domains_suspend(dev_priv,
> > -				    get_suspend_mode(dev_priv, hibernation));
> > +	intel_power_domains_suspend(i915,
> > +				    get_suspend_mode(i915, hibernation));
> >   	ret = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv))
> > -		bxt_enable_dc9(dev_priv);
> > -	else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > -		hsw_enable_pc8(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		ret = vlv_suspend_complete(dev_priv);
> > +	if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915))
> > +		bxt_enable_dc9(i915);
> > +	else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> > +		hsw_enable_pc8(i915);
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		ret = vlv_suspend_complete(i915);
> >   	if (ret) {
> >   		DRM_ERROR("Suspend complete failed: %d\n", ret);
> > -		intel_power_domains_resume(dev_priv);
> > +		intel_power_domains_resume(i915);
> >   		goto out;
> >   	}
> > @@ -2175,13 +2175,13 @@ static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation)
> >   	 * Fujitsu FSC S7110
> >   	 * Acer Aspire 1830T
> >   	 */
> > -	if (!(hibernation && INTEL_GEN(dev_priv) < 6))
> > +	if (!(hibernation && INTEL_GEN(i915) < 6))
> >   		pci_set_power_state(pdev, PCI_D3hot);
> >   out:
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > -	if (!dev_priv->uncore.user_forcewake.count)
> > -		intel_runtime_pm_cleanup(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> > +	if (!i915->uncore.user_forcewake.count)
> > +		intel_runtime_pm_cleanup(i915);
> >   	return ret;
> >   }
> > @@ -2212,24 +2212,24 @@ static int i915_suspend_switcheroo(struct drm_device *dev, pm_message_t state)
> >   static int i915_drm_resume(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int ret;
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > -	intel_sanitize_gt_powersave(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > +	intel_sanitize_gt_powersave(i915);
> > -	i915_gem_sanitize(dev_priv);
> > +	i915_gem_sanitize(i915);
> > -	ret = i915_ggtt_enable_hw(dev_priv);
> > +	ret = i915_ggtt_enable_hw(i915);
> >   	if (ret)
> >   		DRM_ERROR("failed to re-enable GGTT\n");
> > -	intel_csr_ucode_resume(dev_priv);
> > +	intel_csr_ucode_resume(i915);
> > -	i915_restore_state(dev_priv);
> > -	intel_pps_unlock_regs_wa(dev_priv);
> > +	i915_restore_state(i915);
> > +	intel_pps_unlock_regs_wa(i915);
> > -	intel_init_pch_refclk(dev_priv);
> > +	intel_init_pch_refclk(i915);
> >   	/*
> >   	 * Interrupts have to be enabled before any batches are run. If not the
> > @@ -2241,21 +2241,21 @@ static int i915_drm_resume(struct drm_device *dev)
> >   	 * Modeset enabling in intel_modeset_init_hw() also needs working
> >   	 * interrupts.
> >   	 */
> > -	intel_runtime_pm_enable_interrupts(dev_priv);
> > +	intel_runtime_pm_enable_interrupts(i915);
> >   	drm_mode_config_reset(dev);
> > -	i915_gem_resume(dev_priv);
> > +	i915_gem_resume(i915);
> >   	intel_modeset_init_hw(dev);
> > -	intel_init_clock_gating(dev_priv);
> > +	intel_init_clock_gating(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->display.hpd_irq_setup)
> > -		dev_priv->display.hpd_irq_setup(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->display.hpd_irq_setup)
> > +		i915->display.hpd_irq_setup(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	intel_dp_mst_resume(dev_priv);
> > +	intel_dp_mst_resume(i915);
> >   	intel_display_resume(dev);
> > @@ -2267,23 +2267,23 @@ static int i915_drm_resume(struct drm_device *dev)
> >   	 * bother with the tiny race here where we might lose hotplug
> >   	 * notifications.
> >   	 * */
> > -	intel_hpd_init(dev_priv);
> > +	intel_hpd_init(i915);
> > -	intel_opregion_resume(dev_priv);
> > +	intel_opregion_resume(i915);
> >   	intel_fbdev_set_suspend(dev, FBINFO_STATE_RUNNING, false);
> > -	intel_power_domains_enable(dev_priv);
> > +	intel_power_domains_enable(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return 0;
> >   }
> >   static int i915_drm_resume_early(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int ret;
> >   	/*
> > @@ -2330,32 +2330,32 @@ static int i915_drm_resume_early(struct drm_device *dev)
> >   	pci_set_master(pdev);
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		ret = vlv_resume_prepare(dev_priv, false);
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		ret = vlv_resume_prepare(i915, false);
> >   	if (ret)
> >   		DRM_ERROR("Resume prepare failed: %d, continuing anyway\n",
> >   			  ret);
> > -	intel_uncore_resume_early(&dev_priv->uncore);
> > +	intel_uncore_resume_early(&i915->uncore);
> > -	i915_check_and_clear_faults(dev_priv);
> > +	i915_check_and_clear_faults(i915);
> > -	if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv)) {
> > -		gen9_sanitize_dc_state(dev_priv);
> > -		bxt_disable_dc9(dev_priv);
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > -		hsw_disable_pc8(dev_priv);
> > +	if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915)) {
> > +		gen9_sanitize_dc_state(i915);
> > +		bxt_disable_dc9(i915);
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> > +		hsw_disable_pc8(i915);
> >   	}
> > -	intel_uncore_sanitize(dev_priv);
> > +	intel_uncore_sanitize(i915);
> > -	intel_power_domains_resume(dev_priv);
> > +	intel_power_domains_resume(i915);
> > -	intel_gt_sanitize(dev_priv, true);
> > +	intel_gt_sanitize(i915, true);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> > @@ -2540,9 +2540,9 @@ static int i915_pm_restore(struct device *kdev)
> >    * a black-box for the driver. Further investigation is needed to reduce the
> >    * saved/restored registers even further, by following the same 3 criteria.
> >    */
> > -static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> > +static void vlv_save_gunit_s0ix_state(struct drm_i915_private *i915)
> >   {
> > -	struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state;
> > +	struct vlv_s0ix_state *s = &i915->vlv_s0ix_state;
> >   	int i;
> >   	/* GAM 0x4000-0x4770 */
> > @@ -2621,9 +2621,9 @@ static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> >   	 */
> >   }
> > -static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *dev_priv)
> > +static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *i915)
> >   {
> > -	struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state;
> > +	struct vlv_s0ix_state *s = &i915->vlv_s0ix_state;
> >   	u32 val;
> >   	int i;
> > @@ -2732,7 +2732,7 @@ static int vlv_wait_for_pw_status(struct drm_i915_private *i915,
> >   	return ret;
> >   }
> > -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> > +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool force_on)
> >   {
> >   	u32 val;
> >   	int err;
> > @@ -2746,7 +2746,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> >   	if (!force_on)
> >   		return 0;
> > -	err = intel_wait_for_register(&dev_priv->uncore,
> > +	err = intel_wait_for_register(&i915->uncore,
> >   				      VLV_GTLC_SURVIVABILITY_REG,
> >   				      VLV_GFX_CLK_STATUS_BIT,
> >   				      VLV_GFX_CLK_STATUS_BIT,
> > @@ -2758,7 +2758,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on)
> >   	return err;
> >   }
> > -static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow)
> > +static int vlv_allow_gt_wake(struct drm_i915_private *i915, bool allow)
> >   {
> >   	u32 mask;
> >   	u32 val;
> > @@ -2774,14 +2774,14 @@ static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow)
> >   	mask = VLV_GTLC_ALLOWWAKEACK;
> >   	val = allow ? mask : 0;
> > -	err = vlv_wait_for_pw_status(dev_priv, mask, val);
> > +	err = vlv_wait_for_pw_status(i915, mask, val);
> >   	if (err)
> >   		DRM_ERROR("timeout disabling GT waking\n");
> >   	return err;
> >   }
> > -static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv,
> > +static void vlv_wait_for_gt_wells(struct drm_i915_private *i915,
> >   				  bool wait_for_on)
> >   {
> >   	u32 mask;
> > @@ -2797,12 +2797,12 @@ static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv,
> >   	 * This can fail to turn off the rc6 if the GPU is stuck after a failed
> >   	 * reset and we are trying to force the machine to sleep.
> >   	 */
> > -	if (vlv_wait_for_pw_status(dev_priv, mask, val))
> > +	if (vlv_wait_for_pw_status(i915, mask, val))
> >   		DRM_DEBUG_DRIVER("timeout waiting for GT wells to go %s\n",
> >   				 onoff(wait_for_on));
> >   }
> > -static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv)
> > +static void vlv_check_no_gt_access(struct drm_i915_private *i915)
> >   {
> >   	if (!(I915_READ(VLV_GTLC_PW_STATUS) & VLV_GTLC_ALLOWWAKEERR))
> >   		return;
> > @@ -2811,7 +2811,7 @@ static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(VLV_GTLC_PW_STATUS, VLV_GTLC_ALLOWWAKEERR);
> >   }
> > -static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
> > +static int vlv_suspend_complete(struct drm_i915_private *i915)
> >   {
> >   	u32 mask;
> >   	int err;
> > @@ -2820,25 +2820,25 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
> >   	 * Bspec defines the following GT well on flags as debug only, so
> >   	 * don't treat them as hard failures.
> >   	 */
> > -	vlv_wait_for_gt_wells(dev_priv, false);
> > +	vlv_wait_for_gt_wells(i915, false);
> >   	mask = VLV_GTLC_RENDER_CTX_EXISTS | VLV_GTLC_MEDIA_CTX_EXISTS;
> >   	WARN_ON((I915_READ(VLV_GTLC_WAKE_CTRL) & mask) != mask);
> > -	vlv_check_no_gt_access(dev_priv);
> > +	vlv_check_no_gt_access(i915);
> > -	err = vlv_force_gfx_clock(dev_priv, true);
> > +	err = vlv_force_gfx_clock(i915, true);
> >   	if (err)
> >   		goto err1;
> > -	err = vlv_allow_gt_wake(dev_priv, false);
> > +	err = vlv_allow_gt_wake(i915, false);
> >   	if (err)
> >   		goto err2;
> > -	if (!IS_CHERRYVIEW(dev_priv))
> > -		vlv_save_gunit_s0ix_state(dev_priv);
> > +	if (!IS_CHERRYVIEW(i915))
> > +		vlv_save_gunit_s0ix_state(i915);
> > -	err = vlv_force_gfx_clock(dev_priv, false);
> > +	err = vlv_force_gfx_clock(i915, false);
> >   	if (err)
> >   		goto err2;
> > @@ -2846,14 +2846,14 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv)
> >   err2:
> >   	/* For safety always re-enable waking and disable gfx clock forcing */
> > -	vlv_allow_gt_wake(dev_priv, true);
> > +	vlv_allow_gt_wake(i915, true);
> >   err1:
> > -	vlv_force_gfx_clock(dev_priv, false);
> > +	vlv_force_gfx_clock(i915, false);
> >   	return err;
> >   }
> > -static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> > +static int vlv_resume_prepare(struct drm_i915_private *i915,
> >   				bool rpm_resume)
> >   {
> >   	int err;
> > @@ -2864,23 +2864,23 @@ static int vlv_resume_prepare(struct drm_i915_private *dev_priv,
> >   	 * can do at this point. Return the first error code (which will also
> >   	 * leave RPM permanently disabled).
> >   	 */
> > -	ret = vlv_force_gfx_clock(dev_priv, true);
> > +	ret = vlv_force_gfx_clock(i915, true);
> > -	if (!IS_CHERRYVIEW(dev_priv))
> > -		vlv_restore_gunit_s0ix_state(dev_priv);
> > +	if (!IS_CHERRYVIEW(i915))
> > +		vlv_restore_gunit_s0ix_state(i915);
> > -	err = vlv_allow_gt_wake(dev_priv, true);
> > +	err = vlv_allow_gt_wake(i915, true);
> >   	if (!ret)
> >   		ret = err;
> > -	err = vlv_force_gfx_clock(dev_priv, false);
> > +	err = vlv_force_gfx_clock(i915, false);
> >   	if (!ret)
> >   		ret = err;
> > -	vlv_check_no_gt_access(dev_priv);
> > +	vlv_check_no_gt_access(i915);
> >   	if (rpm_resume)
> > -		intel_init_clock_gating(dev_priv);
> > +		intel_init_clock_gating(i915);
> >   	return ret;
> >   }
> > @@ -2889,80 +2889,80 @@ static int intel_runtime_suspend(struct device *kdev)
> >   {
> >   	struct pci_dev *pdev = to_pci_dev(kdev);
> >   	struct drm_device *dev = pci_get_drvdata(pdev);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int ret;
> > -	if (WARN_ON_ONCE(!(dev_priv->gt_pm.rc6.enabled && HAS_RC6(dev_priv))))
> > +	if (WARN_ON_ONCE(!(i915->gt_pm.rc6.enabled && HAS_RC6(i915))))
> >   		return -ENODEV;
> > -	if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv)))
> > +	if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915)))
> >   		return -ENODEV;
> >   	DRM_DEBUG_KMS("Suspending device\n");
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	/*
> >   	 * We are safe here against re-faults, since the fault handler takes
> >   	 * an RPM reference.
> >   	 */
> > -	i915_gem_runtime_suspend(dev_priv);
> > +	i915_gem_runtime_suspend(i915);
> > -	intel_uc_runtime_suspend(dev_priv);
> > +	intel_uc_runtime_suspend(i915);
> > -	intel_runtime_pm_disable_interrupts(dev_priv);
> > +	intel_runtime_pm_disable_interrupts(i915);
> > -	intel_uncore_suspend(&dev_priv->uncore);
> > +	intel_uncore_suspend(&i915->uncore);
> >   	ret = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		icl_display_core_uninit(dev_priv);
> > -		bxt_enable_dc9(dev_priv);
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > -		bxt_display_core_uninit(dev_priv);
> > -		bxt_enable_dc9(dev_priv);
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > -		hsw_enable_pc8(dev_priv);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		ret = vlv_suspend_complete(dev_priv);
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		icl_display_core_uninit(i915);
> > +		bxt_enable_dc9(i915);
> > +	} else if (IS_GEN9_LP(i915)) {
> > +		bxt_display_core_uninit(i915);
> > +		bxt_enable_dc9(i915);
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> > +		hsw_enable_pc8(i915);
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		ret = vlv_suspend_complete(i915);
> >   	}
> >   	if (ret) {
> >   		DRM_ERROR("Runtime suspend failed, disabling it (%d)\n", ret);
> > -		intel_uncore_runtime_resume(&dev_priv->uncore);
> > +		intel_uncore_runtime_resume(&i915->uncore);
> > -		intel_runtime_pm_enable_interrupts(dev_priv);
> > +		intel_runtime_pm_enable_interrupts(i915);
> > -		intel_uc_resume(dev_priv);
> > +		intel_uc_resume(i915);
> > -		i915_gem_init_swizzling(dev_priv);
> > -		i915_gem_restore_fences(dev_priv);
> > +		i915_gem_init_swizzling(i915);
> > +		i915_gem_restore_fences(i915);
> > -		enable_rpm_wakeref_asserts(dev_priv);
> > +		enable_rpm_wakeref_asserts(i915);
> >   		return ret;
> >   	}
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > -	intel_runtime_pm_cleanup(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> > +	intel_runtime_pm_cleanup(i915);
> > -	if (intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore))
> > +	if (intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore))
> >   		DRM_ERROR("Unclaimed access detected prior to suspending\n");
> > -	dev_priv->runtime_pm.suspended = true;
> > +	i915->runtime_pm.suspended = true;
> >   	/*
> >   	 * FIXME: We really should find a document that references the arguments
> >   	 * used below!
> >   	 */
> > -	if (IS_BROADWELL(dev_priv)) {
> > +	if (IS_BROADWELL(i915)) {
> >   		/*
> >   		 * On Broadwell, if we use PCI_D1 the PCH DDI ports will stop
> >   		 * being detected, and the call we do at intel_runtime_resume()
> >   		 * won't be able to restore them. Since PCI_D3hot matches the
> >   		 * actual specification and appears to be working, use it.
> >   		 */
> > -		intel_opregion_notify_adapter(dev_priv, PCI_D3hot);
> > +		intel_opregion_notify_adapter(i915, PCI_D3hot);
> >   	} else {
> >   		/*
> >   		 * current versions of firmware which depend on this opregion
> > @@ -2971,13 +2971,13 @@ static int intel_runtime_suspend(struct device *kdev)
> >   		 * to distinguish it from notifications that might be sent via
> >   		 * the suspend path.
> >   		 */
> > -		intel_opregion_notify_adapter(dev_priv, PCI_D1);
> > +		intel_opregion_notify_adapter(i915, PCI_D1);
> >   	}
> > -	assert_forcewakes_inactive(&dev_priv->uncore);
> > +	assert_forcewakes_inactive(&i915->uncore);
> > -	if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> > -		intel_hpd_poll_init(dev_priv);
> > +	if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915))
> > +		intel_hpd_poll_init(i915);
> >   	DRM_DEBUG_KMS("Device suspended\n");
> >   	return 0;
> > @@ -2987,69 +2987,69 @@ static int intel_runtime_resume(struct device *kdev)
> >   {
> >   	struct pci_dev *pdev = to_pci_dev(kdev);
> >   	struct drm_device *dev = pci_get_drvdata(pdev);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int ret = 0;
> > -	if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv)))
> > +	if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915)))
> >   		return -ENODEV;
> >   	DRM_DEBUG_KMS("Resuming device\n");
> > -	WARN_ON_ONCE(atomic_read(&dev_priv->runtime_pm.wakeref_count));
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	WARN_ON_ONCE(atomic_read(&i915->runtime_pm.wakeref_count));
> > +	disable_rpm_wakeref_asserts(i915);
> > -	intel_opregion_notify_adapter(dev_priv, PCI_D0);
> > -	dev_priv->runtime_pm.suspended = false;
> > -	if (intel_uncore_unclaimed_mmio(&dev_priv->uncore))
> > +	intel_opregion_notify_adapter(i915, PCI_D0);
> > +	i915->runtime_pm.suspended = false;
> > +	if (intel_uncore_unclaimed_mmio(&i915->uncore))
> >   		DRM_DEBUG_DRIVER("Unclaimed access during suspend, bios?\n");
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		bxt_disable_dc9(dev_priv);
> > -		icl_display_core_init(dev_priv, true);
> > -		if (dev_priv->csr.dmc_payload) {
> > -			if (dev_priv->csr.allowed_dc_mask &
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		bxt_disable_dc9(i915);
> > +		icl_display_core_init(i915, true);
> > +		if (i915->csr.dmc_payload) {
> > +			if (i915->csr.allowed_dc_mask &
> >   			    DC_STATE_EN_UPTO_DC6)
> > -				skl_enable_dc6(dev_priv);
> > -			else if (dev_priv->csr.allowed_dc_mask &
> > +				skl_enable_dc6(i915);
> > +			else if (i915->csr.allowed_dc_mask &
> >   				 DC_STATE_EN_UPTO_DC5)
> > -				gen9_enable_dc5(dev_priv);
> > +				gen9_enable_dc5(i915);
> >   		}
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > -		bxt_disable_dc9(dev_priv);
> > -		bxt_display_core_init(dev_priv, true);
> > -		if (dev_priv->csr.dmc_payload &&
> > -		    (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5))
> > -			gen9_enable_dc5(dev_priv);
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > -		hsw_disable_pc8(dev_priv);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		ret = vlv_resume_prepare(dev_priv, true);
> > +	} else if (IS_GEN9_LP(i915)) {
> > +		bxt_disable_dc9(i915);
> > +		bxt_display_core_init(i915, true);
> > +		if (i915->csr.dmc_payload &&
> > +		    (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5))
> > +			gen9_enable_dc5(i915);
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> > +		hsw_disable_pc8(i915);
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		ret = vlv_resume_prepare(i915, true);
> >   	}
> > -	intel_uncore_runtime_resume(&dev_priv->uncore);
> > +	intel_uncore_runtime_resume(&i915->uncore);
> > -	intel_runtime_pm_enable_interrupts(dev_priv);
> > +	intel_runtime_pm_enable_interrupts(i915);
> > -	intel_uc_resume(dev_priv);
> > +	intel_uc_resume(i915);
> >   	/*
> >   	 * No point of rolling back things in case of an error, as the best
> >   	 * we can do is to hope that things will still work (and disable RPM).
> >   	 */
> > -	i915_gem_init_swizzling(dev_priv);
> > -	i915_gem_restore_fences(dev_priv);
> > +	i915_gem_init_swizzling(i915);
> > +	i915_gem_restore_fences(i915);
> >   	/*
> >   	 * On VLV/CHV display interrupts are part of the display
> >   	 * power well, so hpd is reinitialized from there. For
> >   	 * everyone else do it here.
> >   	 */
> > -	if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> > -		intel_hpd_init(dev_priv);
> > +	if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915))
> > +		intel_hpd_init(i915);
> > -	intel_enable_ipc(dev_priv);
> > +	intel_enable_ipc(i915);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	if (ret)
> >   		DRM_ERROR("Runtime resume failed, disabling it (%d)\n", ret);
> > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> > index eb98d285d290..a1adb79ca2c3 100644
> > --- a/drivers/gpu/drm/i915/i915_drv.h
> > +++ b/drivers/gpu/drm/i915/i915_drv.h
> > @@ -208,7 +208,7 @@ struct i915_mm_struct;
> >   struct i915_mmu_object;
> >   struct drm_i915_file_private {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	struct drm_file *file;
> >   	struct {
> > @@ -278,12 +278,12 @@ struct dpll;
> >   struct intel_cdclk_state;
> >   struct drm_i915_display_funcs {
> > -	void (*get_cdclk)(struct drm_i915_private *dev_priv,
> > +	void (*get_cdclk)(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state);
> > -	void (*set_cdclk)(struct drm_i915_private *dev_priv,
> > +	void (*set_cdclk)(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe);
> > -	int (*get_fifo_size)(struct drm_i915_private *dev_priv,
> > +	int (*get_fifo_size)(struct drm_i915_private *i915,
> >   			     enum i9xx_plane_id i9xx_plane);
> >   	int (*compute_pipe_wm)(struct intel_crtc_state *cstate);
> >   	int (*compute_intermediate_wm)(struct intel_crtc_state *newstate);
> > @@ -317,8 +317,8 @@ struct drm_i915_display_funcs {
> >   				    const struct drm_connector_state *old_conn_state);
> >   	void (*fdi_link_train)(struct intel_crtc *crtc,
> >   			       const struct intel_crtc_state *crtc_state);
> > -	void (*init_clock_gating)(struct drm_i915_private *dev_priv);
> > -	void (*hpd_irq_setup)(struct drm_i915_private *dev_priv);
> > +	void (*init_clock_gating)(struct drm_i915_private *i915);
> > +	void (*hpd_irq_setup)(struct drm_i915_private *i915);
> >   	/* clock updates for mode set */
> >   	/* cursor updates */
> >   	/* render clock increase/decrease */
> > @@ -552,7 +552,7 @@ struct intel_gmbus {
> >   	u32 reg0;
> >   	i915_reg_t gpio_reg;
> >   	struct i2c_algo_bit_data bit_algo;
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   };
> >   struct i915_suspend_saved_registers {
> > @@ -641,7 +641,7 @@ struct intel_rps {
> >   	/*
> >   	 * work, interrupts_enabled and pm_iir are protected by
> > -	 * dev_priv->irq_lock
> > +	 * i915->irq_lock
> >   	 */
> >   	struct work_struct work;
> >   	bool interrupts_enabled;
> > @@ -1240,9 +1240,9 @@ struct i915_perf_stream_ops {
> >    */
> >   struct i915_perf_stream {
> >   	/**
> > -	 * @dev_priv: i915 drm device
> > +	 * @i915: i915 drm device
> >   	 */
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	/**
> >   	 * @link: Links the stream into ``&drm_i915_private->streams``
> > @@ -1302,20 +1302,20 @@ struct i915_oa_ops {
> >   	 * @is_valid_b_counter_reg: Validates register's address for
> >   	 * programming boolean counters for a particular platform.
> >   	 */
> > -	bool (*is_valid_b_counter_reg)(struct drm_i915_private *dev_priv,
> > +	bool (*is_valid_b_counter_reg)(struct drm_i915_private *i915,
> >   				       u32 addr);
> >   	/**
> >   	 * @is_valid_mux_reg: Validates register's address for programming mux
> >   	 * for a particular platform.
> >   	 */
> > -	bool (*is_valid_mux_reg)(struct drm_i915_private *dev_priv, u32 addr);
> > +	bool (*is_valid_mux_reg)(struct drm_i915_private *i915, u32 addr);
> >   	/**
> >   	 * @is_valid_flex_reg: Validates register's address for programming
> >   	 * flex EU filtering for a particular platform.
> >   	 */
> > -	bool (*is_valid_flex_reg)(struct drm_i915_private *dev_priv, u32 addr);
> > +	bool (*is_valid_flex_reg)(struct drm_i915_private *i915, u32 addr);
> >   	/**
> >   	 * @enable_metric_set: Selects and applies any MUX configuration to set
> > @@ -1329,7 +1329,7 @@ struct i915_oa_ops {
> >   	 * @disable_metric_set: Remove system constraints associated with using
> >   	 * the OA unit.
> >   	 */
> > -	void (*disable_metric_set)(struct drm_i915_private *dev_priv);
> > +	void (*disable_metric_set)(struct drm_i915_private *i915);
> >   	/**
> >   	 * @oa_enable: Enable periodic sampling
> > @@ -1357,7 +1357,7 @@ struct i915_oa_ops {
> >   	 * handling the OA unit tail pointer race that affects multiple
> >   	 * generations.
> >   	 */
> > -	u32 (*oa_hw_tail_read)(struct drm_i915_private *dev_priv);
> > +	u32 (*oa_hw_tail_read)(struct drm_i915_private *i915);
> >   };
> >   struct intel_cdclk_state {
> > @@ -1750,13 +1750,13 @@ struct drm_i915_private {
> >   		/*
> >   		 * Lock associated with adding/modifying/removing OA configs
> > -		 * in dev_priv->perf.metrics_idr.
> > +		 * in i915->perf.metrics_idr.
> >   		 */
> >   		struct mutex metrics_lock;
> >   		/*
> >   		 * List of dynamic configurations, you need to hold
> > -		 * dev_priv->perf.metrics_lock to access it.
> > +		 * i915->perf.metrics_lock to access it.
> >   		 */
> >   		struct idr metrics_idr;
> > @@ -1772,7 +1772,7 @@ struct drm_i915_private {
> >   			 * The stream currently using the OA unit. If accessed
> >   			 * outside a syscall associated to its file
> >   			 * descriptor, you need to hold
> > -			 * dev_priv->drm.struct_mutex.
> > +			 * i915->drm.struct_mutex.
> >   			 */
> >   			struct i915_perf_stream *exclusive_stream;
> > @@ -2066,15 +2066,15 @@ enum hdmi_force_audio {
> >   	GENMASK(INTEL_FRONTBUFFER_BITS_PER_PIPE * ((pipe) + 1) - 1, \
> >   		INTEL_FRONTBUFFER_BITS_PER_PIPE * (pipe))
> > -#define INTEL_INFO(dev_priv)	(&(dev_priv)->__info)
> > -#define RUNTIME_INFO(dev_priv)	(&(dev_priv)->__runtime)
> > -#define DRIVER_CAPS(dev_priv)	(&(dev_priv)->caps)
> > +#define INTEL_INFO(i915)	(&(i915)->__info)
> > +#define RUNTIME_INFO(i915)	(&(i915)->__runtime)
> > +#define DRIVER_CAPS(i915)	(&(i915)->caps)
> > -#define INTEL_GEN(dev_priv)	(INTEL_INFO(dev_priv)->gen)
> > -#define INTEL_DEVID(dev_priv)	(RUNTIME_INFO(dev_priv)->device_id)
> > +#define INTEL_GEN(i915)	(INTEL_INFO(i915)->gen)
> > +#define INTEL_DEVID(i915)	(RUNTIME_INFO(i915)->device_id)
> >   #define REVID_FOREVER		0xff
> > -#define INTEL_REVID(dev_priv)	((dev_priv)->drm.pdev->revision)
> > +#define INTEL_REVID(i915)	((i915)->drm.pdev->revision)
> >   #define INTEL_GEN_MASK(s, e) ( \
> >   	BUILD_BUG_ON_ZERO(!__builtin_constant_p(s)) + \
> > @@ -2082,12 +2082,12 @@ enum hdmi_force_audio {
> >   	GENMASK((e) - 1, (s) - 1))
> >   /* Returns true if Gen is in inclusive range [Start, End] */
> > -#define IS_GEN_RANGE(dev_priv, s, e) \
> > -	(!!(INTEL_INFO(dev_priv)->gen_mask & INTEL_GEN_MASK((s), (e))))
> > +#define IS_GEN_RANGE(i915, s, e) \
> > +	(!!(INTEL_INFO(i915)->gen_mask & INTEL_GEN_MASK((s), (e))))
> > -#define IS_GEN(dev_priv, n) \
> > +#define IS_GEN(i915, n) \
> >   	(BUILD_BUG_ON_ZERO(!__builtin_constant_p(n)) + \
> > -	 INTEL_INFO(dev_priv)->gen == (n))
> > +	 INTEL_INFO(i915)->gen == (n))
> >   /*
> >    * Return true if revision is in range [since,until] inclusive.
> > @@ -2159,88 +2159,88 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> >   	return ((mask << (msb - pb)) & (mask << (msb - s))) & BIT(msb);
> >   }
> > -#define IS_MOBILE(dev_priv)	(INTEL_INFO(dev_priv)->is_mobile)
> > -
> > -#define IS_I830(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I830)
> > -#define IS_I845G(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I845G)
> > -#define IS_I85X(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I85X)
> > -#define IS_I865G(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I865G)
> > -#define IS_I915G(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I915G)
> > -#define IS_I915GM(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I915GM)
> > -#define IS_I945G(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I945G)
> > -#define IS_I945GM(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I945GM)
> > -#define IS_I965G(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I965G)
> > -#define IS_I965GM(dev_priv)	IS_PLATFORM(dev_priv, INTEL_I965GM)
> > -#define IS_G45(dev_priv)	IS_PLATFORM(dev_priv, INTEL_G45)
> > -#define IS_GM45(dev_priv)	IS_PLATFORM(dev_priv, INTEL_GM45)
> > -#define IS_G4X(dev_priv)	(IS_G45(dev_priv) || IS_GM45(dev_priv))
> > -#define IS_PINEVIEW(dev_priv)	IS_PLATFORM(dev_priv, INTEL_PINEVIEW)
> > -#define IS_G33(dev_priv)	IS_PLATFORM(dev_priv, INTEL_G33)
> > -#define IS_IRONLAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_IRONLAKE)
> > -#define IS_IRONLAKE_M(dev_priv) \
> > -	(IS_PLATFORM(dev_priv, INTEL_IRONLAKE) && IS_MOBILE(dev_priv))
> > -#define IS_IVYBRIDGE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_IVYBRIDGE)
> > -#define IS_IVB_GT1(dev_priv)	(IS_IVYBRIDGE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 1)
> > -#define IS_VALLEYVIEW(dev_priv)	IS_PLATFORM(dev_priv, INTEL_VALLEYVIEW)
> > -#define IS_CHERRYVIEW(dev_priv)	IS_PLATFORM(dev_priv, INTEL_CHERRYVIEW)
> > -#define IS_HASWELL(dev_priv)	IS_PLATFORM(dev_priv, INTEL_HASWELL)
> > -#define IS_BROADWELL(dev_priv)	IS_PLATFORM(dev_priv, INTEL_BROADWELL)
> > -#define IS_SKYLAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_SKYLAKE)
> > -#define IS_BROXTON(dev_priv)	IS_PLATFORM(dev_priv, INTEL_BROXTON)
> > -#define IS_KABYLAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_KABYLAKE)
> > -#define IS_GEMINILAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_GEMINILAKE)
> > -#define IS_COFFEELAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_COFFEELAKE)
> > -#define IS_CANNONLAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_CANNONLAKE)
> > -#define IS_ICELAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_ICELAKE)
> > -#define IS_ELKHARTLAKE(dev_priv)	IS_PLATFORM(dev_priv, INTEL_ELKHARTLAKE)
> > -#define IS_HSW_EARLY_SDV(dev_priv) (IS_HASWELL(dev_priv) && \
> > -				    (INTEL_DEVID(dev_priv) & 0xFF00) == 0x0C00)
> > -#define IS_BDW_ULT(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT)
> > -#define IS_BDW_ULX(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX)
> > -#define IS_BDW_GT3(dev_priv)	(IS_BROADWELL(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 3)
> > -#define IS_HSW_ULT(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT)
> > -#define IS_HSW_GT3(dev_priv)	(IS_HASWELL(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 3)
> > -#define IS_HSW_GT1(dev_priv)	(IS_HASWELL(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 1)
> > +#define IS_MOBILE(i915)	(INTEL_INFO(i915)->is_mobile)
> > +
> > +#define IS_I830(i915)	IS_PLATFORM(i915, INTEL_I830)
> > +#define IS_I845G(i915)	IS_PLATFORM(i915, INTEL_I845G)
> > +#define IS_I85X(i915)	IS_PLATFORM(i915, INTEL_I85X)
> > +#define IS_I865G(i915)	IS_PLATFORM(i915, INTEL_I865G)
> > +#define IS_I915G(i915)	IS_PLATFORM(i915, INTEL_I915G)
> > +#define IS_I915GM(i915)	IS_PLATFORM(i915, INTEL_I915GM)
> > +#define IS_I945G(i915)	IS_PLATFORM(i915, INTEL_I945G)
> > +#define IS_I945GM(i915)	IS_PLATFORM(i915, INTEL_I945GM)
> > +#define IS_I965G(i915)	IS_PLATFORM(i915, INTEL_I965G)
> > +#define IS_I965GM(i915)	IS_PLATFORM(i915, INTEL_I965GM)
> > +#define IS_G45(i915)	IS_PLATFORM(i915, INTEL_G45)
> > +#define IS_GM45(i915)	IS_PLATFORM(i915, INTEL_GM45)
> > +#define IS_G4X(i915)	(IS_G45(i915) || IS_GM45(i915))
> > +#define IS_PINEVIEW(i915)	IS_PLATFORM(i915, INTEL_PINEVIEW)
> > +#define IS_G33(i915)	IS_PLATFORM(i915, INTEL_G33)
> > +#define IS_IRONLAKE(i915)	IS_PLATFORM(i915, INTEL_IRONLAKE)
> > +#define IS_IRONLAKE_M(i915) \
> > +	(IS_PLATFORM(i915, INTEL_IRONLAKE) && IS_MOBILE(i915))
> > +#define IS_IVYBRIDGE(i915)	IS_PLATFORM(i915, INTEL_IVYBRIDGE)
> > +#define IS_IVB_GT1(i915)	(IS_IVYBRIDGE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 1)
> > +#define IS_VALLEYVIEW(i915)	IS_PLATFORM(i915, INTEL_VALLEYVIEW)
> > +#define IS_CHERRYVIEW(i915)	IS_PLATFORM(i915, INTEL_CHERRYVIEW)
> > +#define IS_HASWELL(i915)	IS_PLATFORM(i915, INTEL_HASWELL)
> > +#define IS_BROADWELL(i915)	IS_PLATFORM(i915, INTEL_BROADWELL)
> > +#define IS_SKYLAKE(i915)	IS_PLATFORM(i915, INTEL_SKYLAKE)
> > +#define IS_BROXTON(i915)	IS_PLATFORM(i915, INTEL_BROXTON)
> > +#define IS_KABYLAKE(i915)	IS_PLATFORM(i915, INTEL_KABYLAKE)
> > +#define IS_GEMINILAKE(i915)	IS_PLATFORM(i915, INTEL_GEMINILAKE)
> > +#define IS_COFFEELAKE(i915)	IS_PLATFORM(i915, INTEL_COFFEELAKE)
> > +#define IS_CANNONLAKE(i915)	IS_PLATFORM(i915, INTEL_CANNONLAKE)
> > +#define IS_ICELAKE(i915)	IS_PLATFORM(i915, INTEL_ICELAKE)
> > +#define IS_ELKHARTLAKE(i915)	IS_PLATFORM(i915, INTEL_ELKHARTLAKE)
> > +#define IS_HSW_EARLY_SDV(i915) (IS_HASWELL(i915) && \
> > +				    (INTEL_DEVID(i915) & 0xFF00) == 0x0C00)
> > +#define IS_BDW_ULT(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT)
> > +#define IS_BDW_ULX(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX)
> > +#define IS_BDW_GT3(i915)	(IS_BROADWELL(i915) && \
> > +				 INTEL_INFO(i915)->gt == 3)
> > +#define IS_HSW_ULT(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT)
> > +#define IS_HSW_GT3(i915)	(IS_HASWELL(i915) && \
> > +				 INTEL_INFO(i915)->gt == 3)
> > +#define IS_HSW_GT1(i915)	(IS_HASWELL(i915) && \
> > +				 INTEL_INFO(i915)->gt == 1)
> >   /* ULX machines are also considered ULT. */
> > -#define IS_HSW_ULX(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX)
> > -#define IS_SKL_ULT(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT)
> > -#define IS_SKL_ULX(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX)
> > -#define IS_KBL_ULT(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT)
> > -#define IS_KBL_ULX(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX)
> > -#define IS_SKL_GT2(dev_priv)	(IS_SKYLAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 2)
> > -#define IS_SKL_GT3(dev_priv)	(IS_SKYLAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 3)
> > -#define IS_SKL_GT4(dev_priv)	(IS_SKYLAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 4)
> > -#define IS_KBL_GT2(dev_priv)	(IS_KABYLAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 2)
> > -#define IS_KBL_GT3(dev_priv)	(IS_KABYLAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 3)
> > -#define IS_CFL_ULT(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT)
> > -#define IS_CFL_ULX(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX)
> > -#define IS_CFL_GT2(dev_priv)	(IS_COFFEELAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 2)
> > -#define IS_CFL_GT3(dev_priv)	(IS_COFFEELAKE(dev_priv) && \
> > -				 INTEL_INFO(dev_priv)->gt == 3)
> > -#define IS_CNL_WITH_PORT_F(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF)
> > -#define IS_ICL_WITH_PORT_F(dev_priv) \
> > -	IS_SUBPLATFORM(dev_priv, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF)
> > +#define IS_HSW_ULX(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX)
> > +#define IS_SKL_ULT(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT)
> > +#define IS_SKL_ULX(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX)
> > +#define IS_KBL_ULT(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT)
> > +#define IS_KBL_ULX(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX)
> > +#define IS_SKL_GT2(i915)	(IS_SKYLAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 2)
> > +#define IS_SKL_GT3(i915)	(IS_SKYLAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 3)
> > +#define IS_SKL_GT4(i915)	(IS_SKYLAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 4)
> > +#define IS_KBL_GT2(i915)	(IS_KABYLAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 2)
> > +#define IS_KBL_GT3(i915)	(IS_KABYLAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 3)
> > +#define IS_CFL_ULT(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT)
> > +#define IS_CFL_ULX(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX)
> > +#define IS_CFL_GT2(i915)	(IS_COFFEELAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 2)
> > +#define IS_CFL_GT3(i915)	(IS_COFFEELAKE(i915) && \
> > +				 INTEL_INFO(i915)->gt == 3)
> > +#define IS_CNL_WITH_PORT_F(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF)
> > +#define IS_ICL_WITH_PORT_F(i915) \
> > +	IS_SUBPLATFORM(i915, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF)
> >   #define SKL_REVID_A0		0x0
> >   #define SKL_REVID_B0		0x1
> > @@ -2259,8 +2259,8 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> >   #define BXT_REVID_B_LAST	0x8
> >   #define BXT_REVID_C0		0x9
> > -#define IS_BXT_REVID(dev_priv, since, until) \
> > -	(IS_BROXTON(dev_priv) && IS_REVID(dev_priv, since, until))
> > +#define IS_BXT_REVID(i915, since, until) \
> > +	(IS_BROXTON(i915) && IS_REVID(i915, since, until))
> >   #define KBL_REVID_A0		0x0
> >   #define KBL_REVID_B0		0x1
> > @@ -2268,14 +2268,14 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> >   #define KBL_REVID_D0		0x3
> >   #define KBL_REVID_E0		0x4
> > -#define IS_KBL_REVID(dev_priv, since, until) \
> > -	(IS_KABYLAKE(dev_priv) && IS_REVID(dev_priv, since, until))
> > +#define IS_KBL_REVID(i915, since, until) \
> > +	(IS_KABYLAKE(i915) && IS_REVID(i915, since, until))
> >   #define GLK_REVID_A0		0x0
> >   #define GLK_REVID_A1		0x1
> > -#define IS_GLK_REVID(dev_priv, since, until) \
> > -	(IS_GEMINILAKE(dev_priv) && IS_REVID(dev_priv, since, until))
> > +#define IS_GLK_REVID(i915, since, until) \
> > +	(IS_GEMINILAKE(i915) && IS_REVID(i915, since, until))
> >   #define CNL_REVID_A0		0x0
> >   #define CNL_REVID_B0		0x1
> > @@ -2293,122 +2293,122 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> >   #define IS_ICL_REVID(p, since, until) \
> >   	(IS_ICELAKE(p) && IS_REVID(p, since, until))
> > -#define IS_LP(dev_priv)	(INTEL_INFO(dev_priv)->is_lp)
> > -#define IS_GEN9_LP(dev_priv)	(IS_GEN(dev_priv, 9) && IS_LP(dev_priv))
> > -#define IS_GEN9_BC(dev_priv)	(IS_GEN(dev_priv, 9) && !IS_LP(dev_priv))
> > +#define IS_LP(i915)	(INTEL_INFO(i915)->is_lp)
> > +#define IS_GEN9_LP(i915)	(IS_GEN(i915, 9) && IS_LP(i915))
> > +#define IS_GEN9_BC(i915)	(IS_GEN(i915, 9) && !IS_LP(i915))
> > -#define HAS_ENGINE(dev_priv, id) (INTEL_INFO(dev_priv)->engine_mask & BIT(id))
> > +#define HAS_ENGINE(i915, id) (INTEL_INFO(i915)->engine_mask & BIT(id))
> > -#define ENGINE_INSTANCES_MASK(dev_priv, first, count) ({		\
> > +#define ENGINE_INSTANCES_MASK(i915, first, count) ({		\
> >   	unsigned int first__ = (first);					\
> >   	unsigned int count__ = (count);					\
> > -	(INTEL_INFO(dev_priv)->engine_mask &				\
> > +	(INTEL_INFO(i915)->engine_mask &				\
> >   	 GENMASK(first__ + count__ - 1, first__)) >> first__;		\
> >   })
> > -#define VDBOX_MASK(dev_priv) \
> > -	ENGINE_INSTANCES_MASK(dev_priv, VCS0, I915_MAX_VCS)
> > -#define VEBOX_MASK(dev_priv) \
> > -	ENGINE_INSTANCES_MASK(dev_priv, VECS0, I915_MAX_VECS)
> > -
> > -#define HAS_LLC(dev_priv)	(INTEL_INFO(dev_priv)->has_llc)
> > -#define HAS_SNOOP(dev_priv)	(INTEL_INFO(dev_priv)->has_snoop)
> > -#define HAS_EDRAM(dev_priv)	((dev_priv)->edram_size_mb)
> > -#define HAS_WT(dev_priv)	((IS_HASWELL(dev_priv) || \
> > -				 IS_BROADWELL(dev_priv)) && HAS_EDRAM(dev_priv))
> > -
> > -#define HWS_NEEDS_PHYSICAL(dev_priv)	(INTEL_INFO(dev_priv)->hws_needs_physical)
> > -
> > -#define HAS_LOGICAL_RING_CONTEXTS(dev_priv) \
> > -		(INTEL_INFO(dev_priv)->has_logical_ring_contexts)
> > -#define HAS_LOGICAL_RING_ELSQ(dev_priv) \
> > -		(INTEL_INFO(dev_priv)->has_logical_ring_elsq)
> > -#define HAS_LOGICAL_RING_PREEMPTION(dev_priv) \
> > -		(INTEL_INFO(dev_priv)->has_logical_ring_preemption)
> > -
> > -#define HAS_EXECLISTS(dev_priv) HAS_LOGICAL_RING_CONTEXTS(dev_priv)
> > -
> > -#define INTEL_PPGTT(dev_priv) (INTEL_INFO(dev_priv)->ppgtt_type)
> > -#define HAS_PPGTT(dev_priv) \
> > -	(INTEL_PPGTT(dev_priv) != INTEL_PPGTT_NONE)
> > -#define HAS_FULL_PPGTT(dev_priv) \
> > -	(INTEL_PPGTT(dev_priv) >= INTEL_PPGTT_FULL)
> > -
> > -#define HAS_PAGE_SIZES(dev_priv, sizes) ({ \
> > +#define VDBOX_MASK(i915) \
> > +	ENGINE_INSTANCES_MASK(i915, VCS0, I915_MAX_VCS)
> > +#define VEBOX_MASK(i915) \
> > +	ENGINE_INSTANCES_MASK(i915, VECS0, I915_MAX_VECS)
> > +
> > +#define HAS_LLC(i915)	(INTEL_INFO(i915)->has_llc)
> > +#define HAS_SNOOP(i915)	(INTEL_INFO(i915)->has_snoop)
> > +#define HAS_EDRAM(i915)	((i915)->edram_size_mb)
> > +#define HAS_WT(i915)	((IS_HASWELL(i915) || \
> > +				 IS_BROADWELL(i915)) && HAS_EDRAM(i915))
> > +
> > +#define HWS_NEEDS_PHYSICAL(i915)	(INTEL_INFO(i915)->hws_needs_physical)
> > +
> > +#define HAS_LOGICAL_RING_CONTEXTS(i915) \
> > +		(INTEL_INFO(i915)->has_logical_ring_contexts)
> > +#define HAS_LOGICAL_RING_ELSQ(i915) \
> > +		(INTEL_INFO(i915)->has_logical_ring_elsq)
> > +#define HAS_LOGICAL_RING_PREEMPTION(i915) \
> > +		(INTEL_INFO(i915)->has_logical_ring_preemption)
> > +
> > +#define HAS_EXECLISTS(i915) HAS_LOGICAL_RING_CONTEXTS(i915)
> > +
> > +#define INTEL_PPGTT(i915) (INTEL_INFO(i915)->ppgtt_type)
> > +#define HAS_PPGTT(i915) \
> > +	(INTEL_PPGTT(i915) != INTEL_PPGTT_NONE)
> > +#define HAS_FULL_PPGTT(i915) \
> > +	(INTEL_PPGTT(i915) >= INTEL_PPGTT_FULL)
> > +
> > +#define HAS_PAGE_SIZES(i915, sizes) ({ \
> >   	GEM_BUG_ON((sizes) == 0); \
> > -	((sizes) & ~INTEL_INFO(dev_priv)->page_sizes) == 0; \
> > +	((sizes) & ~INTEL_INFO(i915)->page_sizes) == 0; \
> >   })
> > -#define HAS_OVERLAY(dev_priv)		 (INTEL_INFO(dev_priv)->display.has_overlay)
> > -#define OVERLAY_NEEDS_PHYSICAL(dev_priv) \
> > -		(INTEL_INFO(dev_priv)->display.overlay_needs_physical)
> > +#define HAS_OVERLAY(i915)		 (INTEL_INFO(i915)->display.has_overlay)
> > +#define OVERLAY_NEEDS_PHYSICAL(i915) \
> > +		(INTEL_INFO(i915)->display.overlay_needs_physical)
> >   /* Early gen2 have a totally busted CS tlb and require pinned batches. */
> > -#define HAS_BROKEN_CS_TLB(dev_priv)	(IS_I830(dev_priv) || IS_I845G(dev_priv))
> > +#define HAS_BROKEN_CS_TLB(i915)	(IS_I830(i915) || IS_I845G(i915))
> >   /* WaRsDisableCoarsePowerGating:skl,cnl */
> > -#define NEEDS_WaRsDisableCoarsePowerGating(dev_priv) \
> > -	(IS_CANNONLAKE(dev_priv) || \
> > -	 IS_SKL_GT3(dev_priv) || IS_SKL_GT4(dev_priv))
> > +#define NEEDS_WaRsDisableCoarsePowerGating(i915) \
> > +	(IS_CANNONLAKE(i915) || \
> > +	 IS_SKL_GT3(i915) || IS_SKL_GT4(i915))
> > -#define HAS_GMBUS_IRQ(dev_priv) (INTEL_GEN(dev_priv) >= 4)
> > -#define HAS_GMBUS_BURST_READ(dev_priv) (INTEL_GEN(dev_priv) >= 10 || \
> > -					IS_GEMINILAKE(dev_priv) || \
> > -					IS_KABYLAKE(dev_priv))
> > +#define HAS_GMBUS_IRQ(i915) (INTEL_GEN(i915) >= 4)
> > +#define HAS_GMBUS_BURST_READ(i915) (INTEL_GEN(i915) >= 10 || \
> > +					IS_GEMINILAKE(i915) || \
> > +					IS_KABYLAKE(i915))
> >   /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte
> >    * rows, which changed the alignment requirements and fence programming.
> >    */
> > -#define HAS_128_BYTE_Y_TILING(dev_priv) (!IS_GEN(dev_priv, 2) && \
> > -					 !(IS_I915G(dev_priv) || \
> > -					 IS_I915GM(dev_priv)))
> > -#define SUPPORTS_TV(dev_priv)		(INTEL_INFO(dev_priv)->display.supports_tv)
> > -#define I915_HAS_HOTPLUG(dev_priv)	(INTEL_INFO(dev_priv)->display.has_hotplug)
> > +#define HAS_128_BYTE_Y_TILING(i915) (!IS_GEN(i915, 2) && \
> > +					 !(IS_I915G(i915) || \
> > +					 IS_I915GM(i915)))
> > +#define SUPPORTS_TV(i915)		(INTEL_INFO(i915)->display.supports_tv)
> > +#define I915_HAS_HOTPLUG(i915)	(INTEL_INFO(i915)->display.has_hotplug)
> > -#define HAS_FW_BLC(dev_priv) 	(INTEL_GEN(dev_priv) > 2)
> > -#define HAS_FBC(dev_priv)	(INTEL_INFO(dev_priv)->display.has_fbc)
> > -#define HAS_CUR_FBC(dev_priv)	(!HAS_GMCH(dev_priv) && INTEL_GEN(dev_priv) >= 7)
> > +#define HAS_FW_BLC(i915) 	(INTEL_GEN(i915) > 2)
> > +#define HAS_FBC(i915)	(INTEL_INFO(i915)->display.has_fbc)
> > +#define HAS_CUR_FBC(i915)	(!HAS_GMCH(i915) && INTEL_GEN(i915) >= 7)
> > -#define HAS_IPS(dev_priv)	(IS_HSW_ULT(dev_priv) || IS_BROADWELL(dev_priv))
> > +#define HAS_IPS(i915)	(IS_HSW_ULT(i915) || IS_BROADWELL(i915))
> > -#define HAS_DP_MST(dev_priv)	(INTEL_INFO(dev_priv)->display.has_dp_mst)
> > +#define HAS_DP_MST(i915)	(INTEL_INFO(i915)->display.has_dp_mst)
> > -#define HAS_DDI(dev_priv)		 (INTEL_INFO(dev_priv)->display.has_ddi)
> > -#define HAS_FPGA_DBG_UNCLAIMED(dev_priv) (INTEL_INFO(dev_priv)->has_fpga_dbg)
> > -#define HAS_PSR(dev_priv)		 (INTEL_INFO(dev_priv)->display.has_psr)
> > -#define HAS_TRANSCODER_EDP(dev_priv)	 (INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_EDP] != 0)
> > +#define HAS_DDI(i915)		 (INTEL_INFO(i915)->display.has_ddi)
> > +#define HAS_FPGA_DBG_UNCLAIMED(i915) (INTEL_INFO(i915)->has_fpga_dbg)
> > +#define HAS_PSR(i915)		 (INTEL_INFO(i915)->display.has_psr)
> > +#define HAS_TRANSCODER_EDP(i915)	 (INTEL_INFO(i915)->trans_offsets[TRANSCODER_EDP] != 0)
> > -#define HAS_RC6(dev_priv)		 (INTEL_INFO(dev_priv)->has_rc6)
> > -#define HAS_RC6p(dev_priv)		 (INTEL_INFO(dev_priv)->has_rc6p)
> > -#define HAS_RC6pp(dev_priv)		 (false) /* HW was never validated */
> > +#define HAS_RC6(i915)		 (INTEL_INFO(i915)->has_rc6)
> > +#define HAS_RC6p(i915)		 (INTEL_INFO(i915)->has_rc6p)
> > +#define HAS_RC6pp(i915)		 (false) /* HW was never validated */
> > -#define HAS_RPS(dev_priv)	(INTEL_INFO(dev_priv)->has_rps)
> > +#define HAS_RPS(i915)	(INTEL_INFO(i915)->has_rps)
> > -#define HAS_CSR(dev_priv)	(INTEL_INFO(dev_priv)->display.has_csr)
> > +#define HAS_CSR(i915)	(INTEL_INFO(i915)->display.has_csr)
> > -#define HAS_RUNTIME_PM(dev_priv) (INTEL_INFO(dev_priv)->has_runtime_pm)
> > -#define HAS_64BIT_RELOC(dev_priv) (INTEL_INFO(dev_priv)->has_64bit_reloc)
> > +#define HAS_RUNTIME_PM(i915) (INTEL_INFO(i915)->has_runtime_pm)
> > +#define HAS_64BIT_RELOC(i915) (INTEL_INFO(i915)->has_64bit_reloc)
> > -#define HAS_IPC(dev_priv)		 (INTEL_INFO(dev_priv)->display.has_ipc)
> > +#define HAS_IPC(i915)		 (INTEL_INFO(i915)->display.has_ipc)
> >   /*
> >    * For now, anything with a GuC requires uCode loading, and then supports
> >    * command submission once loaded. But these are logically independent
> >    * properties, so we have separate macros to test them.
> >    */
> > -#define HAS_GUC(dev_priv)	(INTEL_INFO(dev_priv)->has_guc)
> > -#define HAS_GUC_UCODE(dev_priv)	(HAS_GUC(dev_priv))
> > -#define HAS_GUC_SCHED(dev_priv)	(HAS_GUC(dev_priv))
> > +#define HAS_GUC(i915)	(INTEL_INFO(i915)->has_guc)
> > +#define HAS_GUC_UCODE(i915)	(HAS_GUC(i915))
> > +#define HAS_GUC_SCHED(i915)	(HAS_GUC(i915))
> >   /* For now, anything with a GuC has also HuC */
> > -#define HAS_HUC(dev_priv)	(HAS_GUC(dev_priv))
> > -#define HAS_HUC_UCODE(dev_priv)	(HAS_GUC(dev_priv))
> > +#define HAS_HUC(i915)	(HAS_GUC(i915))
> > +#define HAS_HUC_UCODE(i915)	(HAS_GUC(i915))
> >   /* Having a GuC is not the same as using a GuC */
> > -#define USES_GUC(dev_priv)		intel_uc_is_using_guc(dev_priv)
> > -#define USES_GUC_SUBMISSION(dev_priv)	intel_uc_is_using_guc_submission(dev_priv)
> > -#define USES_HUC(dev_priv)		intel_uc_is_using_huc(dev_priv)
> > +#define USES_GUC(i915)		intel_uc_is_using_guc(i915)
> > +#define USES_GUC_SUBMISSION(i915)	intel_uc_is_using_guc_submission(i915)
> > +#define USES_HUC(i915)		intel_uc_is_using_huc(i915)
> > -#define HAS_POOLED_EU(dev_priv)	(INTEL_INFO(dev_priv)->has_pooled_eu)
> > +#define HAS_POOLED_EU(i915)	(INTEL_INFO(i915)->has_pooled_eu)
> >   #define INTEL_PCH_DEVICE_ID_MASK		0xff80
> >   #define INTEL_PCH_IBX_DEVICE_ID_TYPE		0x3b00
> > @@ -2429,36 +2429,36 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915,
> >   #define INTEL_PCH_P3X_DEVICE_ID_TYPE		0x7000
> >   #define INTEL_PCH_QEMU_DEVICE_ID_TYPE		0x2900 /* qemu q35 has 2918 */
> > -#define INTEL_PCH_TYPE(dev_priv) ((dev_priv)->pch_type)
> > -#define INTEL_PCH_ID(dev_priv) ((dev_priv)->pch_id)
> > -#define HAS_PCH_ICP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_ICP)
> > -#define HAS_PCH_CNP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CNP)
> > -#define HAS_PCH_SPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_SPT)
> > -#define HAS_PCH_LPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_LPT)
> > -#define HAS_PCH_LPT_LP(dev_priv) \
> > -	(INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \
> > -	 INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE)
> > -#define HAS_PCH_LPT_H(dev_priv) \
> > -	(INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \
> > -	 INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_DEVICE_ID_TYPE)
> > -#define HAS_PCH_CPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CPT)
> > -#define HAS_PCH_IBX(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_IBX)
> > -#define HAS_PCH_NOP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_NOP)
> > -#define HAS_PCH_SPLIT(dev_priv) (INTEL_PCH_TYPE(dev_priv) != PCH_NONE)
> > -
> > -#define HAS_GMCH(dev_priv) (INTEL_INFO(dev_priv)->display.has_gmch)
> > -
> > -#define HAS_LSPCON(dev_priv) (INTEL_GEN(dev_priv) >= 9)
> > +#define INTEL_PCH_TYPE(i915) ((i915)->pch_type)
> > +#define INTEL_PCH_ID(i915) ((i915)->pch_id)
> > +#define HAS_PCH_ICP(i915) (INTEL_PCH_TYPE(i915) == PCH_ICP)
> > +#define HAS_PCH_CNP(i915) (INTEL_PCH_TYPE(i915) == PCH_CNP)
> > +#define HAS_PCH_SPT(i915) (INTEL_PCH_TYPE(i915) == PCH_SPT)
> > +#define HAS_PCH_LPT(i915) (INTEL_PCH_TYPE(i915) == PCH_LPT)
> > +#define HAS_PCH_LPT_LP(i915) \
> > +	(INTEL_PCH_ID(i915) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \
> > +	 INTEL_PCH_ID(i915) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE)
> > +#define HAS_PCH_LPT_H(i915) \
> > +	(INTEL_PCH_ID(i915) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \
> > +	 INTEL_PCH_ID(i915) == INTEL_PCH_WPT_DEVICE_ID_TYPE)
> > +#define HAS_PCH_CPT(i915) (INTEL_PCH_TYPE(i915) == PCH_CPT)
> > +#define HAS_PCH_IBX(i915) (INTEL_PCH_TYPE(i915) == PCH_IBX)
> > +#define HAS_PCH_NOP(i915) (INTEL_PCH_TYPE(i915) == PCH_NOP)
> > +#define HAS_PCH_SPLIT(i915) (INTEL_PCH_TYPE(i915) != PCH_NONE)
> > +
> > +#define HAS_GMCH(i915) (INTEL_INFO(i915)->display.has_gmch)
> > +
> > +#define HAS_LSPCON(i915) (INTEL_GEN(i915) >= 9)
> >   /* DPF == dynamic parity feature */
> > -#define HAS_L3_DPF(dev_priv) (INTEL_INFO(dev_priv)->has_l3_dpf)
> > -#define NUM_L3_SLICES(dev_priv) (IS_HSW_GT3(dev_priv) ? \
> > -				 2 : HAS_L3_DPF(dev_priv))
> > +#define HAS_L3_DPF(i915) (INTEL_INFO(i915)->has_l3_dpf)
> > +#define NUM_L3_SLICES(i915) (IS_HSW_GT3(i915) ? \
> > +				 2 : HAS_L3_DPF(i915))
> >   #define GT_FREQUENCY_MULTIPLIER 50
> >   #define GEN9_FREQ_SCALER 3
> > -#define HAS_DISPLAY(dev_priv) (INTEL_INFO(dev_priv)->num_pipes > 0)
> > +#define HAS_DISPLAY(i915) (INTEL_INFO(i915)->num_pipes > 0)
> >   #include "i915_trace.h"
> > @@ -2471,24 +2471,24 @@ static inline bool intel_vtd_active(void)
> >   	return false;
> >   }
> > -static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *dev_priv)
> > +static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *i915)
> >   {
> > -	return INTEL_GEN(dev_priv) >= 6 && intel_vtd_active();
> > +	return INTEL_GEN(i915) >= 6 && intel_vtd_active();
> >   }
> >   static inline bool
> > -intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *dev_priv)
> > +intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *i915)
> >   {
> > -	return IS_BROXTON(dev_priv) && intel_vtd_active();
> > +	return IS_BROXTON(i915) && intel_vtd_active();
> >   }
> >   /* i915_drv.c */
> >   void __printf(3, 4)
> > -__i915_printk(struct drm_i915_private *dev_priv, const char *level,
> > +__i915_printk(struct drm_i915_private *i915, const char *level,
> >   	      const char *fmt, ...);
> > -#define i915_report_error(dev_priv, fmt, ...)				   \
> > -	__i915_printk(dev_priv, KERN_ERR, fmt, ##__VA_ARGS__)
> > +#define i915_report_error(i915, fmt, ...)				   \
> > +	__i915_printk(i915, KERN_ERR, fmt, ##__VA_ARGS__)
> >   #ifdef CONFIG_COMPAT
> >   extern long i915_compat_ioctl(struct file *filp, unsigned int cmd,
> > @@ -2503,12 +2503,12 @@ extern int i915_driver_load(struct pci_dev *pdev,
> >   extern void i915_driver_unload(struct drm_device *dev);
> >   extern void intel_engine_init_hangcheck(struct intel_engine_cs *engine);
> > -extern void intel_hangcheck_init(struct drm_i915_private *dev_priv);
> > -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool on);
> > +extern void intel_hangcheck_init(struct drm_i915_private *i915);
> > +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool on);
> > -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv);
> > +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915);
> > -static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv)
> > +static inline void i915_queue_hangcheck(struct drm_i915_private *i915)
> >   {
> >   	unsigned long delay;
> > @@ -2522,28 +2522,28 @@ static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv)
> >   	delay = round_jiffies_up_relative(DRM_I915_HANGCHECK_JIFFIES);
> >   	queue_delayed_work(system_long_wq,
> > -			   &dev_priv->gpu_error.hangcheck_work, delay);
> > +			   &i915->gpu_error.hangcheck_work, delay);
> >   }
> > -static inline bool intel_gvt_active(struct drm_i915_private *dev_priv)
> > +static inline bool intel_gvt_active(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->gvt;
> > +	return i915->gvt;
> >   }
> > -static inline bool intel_vgpu_active(struct drm_i915_private *dev_priv)
> > +static inline bool intel_vgpu_active(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->vgpu.active;
> > +	return i915->vgpu.active;
> >   }
> >   /* i915_gem.c */
> > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv);
> > -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv);
> > +int i915_gem_init_userptr(struct drm_i915_private *i915);
> > +void i915_gem_cleanup_userptr(struct drm_i915_private *i915);
> >   void i915_gem_sanitize(struct drm_i915_private *i915);
> > -int i915_gem_init_early(struct drm_i915_private *dev_priv);
> > -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv);
> > -void i915_gem_load_init_fences(struct drm_i915_private *dev_priv);
> > -int i915_gem_freeze(struct drm_i915_private *dev_priv);
> > -int i915_gem_freeze_late(struct drm_i915_private *dev_priv);
> > +int i915_gem_init_early(struct drm_i915_private *i915);
> > +void i915_gem_cleanup_early(struct drm_i915_private *i915);
> > +void i915_gem_load_init_fences(struct drm_i915_private *i915);
> > +int i915_gem_freeze(struct drm_i915_private *i915);
> > +int i915_gem_freeze_late(struct drm_i915_private *i915);
> >   static inline void i915_gem_drain_freed_objects(struct drm_i915_private *i915)
> >   {
> > @@ -2591,7 +2591,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> >   int i915_gem_object_unbind(struct drm_i915_gem_object *obj);
> > -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv);
> > +void i915_gem_runtime_suspend(struct drm_i915_private *i915);
> >   static inline int __must_check
> >   i915_mutex_lock_interruptible(struct drm_device *dev)
> > @@ -2633,20 +2633,20 @@ static inline u32 i915_reset_engine_count(struct i915_gpu_error *error,
> >   	return READ_ONCE(error->reset_engine_count[engine->id]);
> >   }
> > -void i915_gem_set_wedged(struct drm_i915_private *dev_priv);
> > -bool i915_gem_unset_wedged(struct drm_i915_private *dev_priv);
> > +void i915_gem_set_wedged(struct drm_i915_private *i915);
> > +bool i915_gem_unset_wedged(struct drm_i915_private *i915);
> >   void i915_gem_init_mmio(struct drm_i915_private *i915);
> > -int __must_check i915_gem_init(struct drm_i915_private *dev_priv);
> > -int __must_check i915_gem_init_hw(struct drm_i915_private *dev_priv);
> > -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv);
> > -void i915_gem_fini_hw(struct drm_i915_private *dev_priv);
> > -void i915_gem_fini(struct drm_i915_private *dev_priv);
> > -int i915_gem_wait_for_idle(struct drm_i915_private *dev_priv,
> > +int __must_check i915_gem_init(struct drm_i915_private *i915);
> > +int __must_check i915_gem_init_hw(struct drm_i915_private *i915);
> > +void i915_gem_init_swizzling(struct drm_i915_private *i915);
> > +void i915_gem_fini_hw(struct drm_i915_private *i915);
> > +void i915_gem_fini(struct drm_i915_private *i915);
> > +int i915_gem_wait_for_idle(struct drm_i915_private *i915,
> >   			   unsigned int flags, long timeout);
> > -void i915_gem_suspend(struct drm_i915_private *dev_priv);
> > -void i915_gem_suspend_late(struct drm_i915_private *dev_priv);
> > -void i915_gem_resume(struct drm_i915_private *dev_priv);
> > +void i915_gem_suspend(struct drm_i915_private *i915);
> > +void i915_gem_suspend_late(struct drm_i915_private *i915);
> > +void i915_gem_resume(struct drm_i915_private *i915);
> >   vm_fault_t i915_gem_fault(struct vm_fault *vmf);
> >   int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file);
> > @@ -2663,12 +2663,12 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev,
> >   /* i915_gem_fence_reg.c */
> >   struct drm_i915_fence_reg *
> > -i915_reserve_fence(struct drm_i915_private *dev_priv);
> > +i915_reserve_fence(struct drm_i915_private *i915);
> >   void i915_unreserve_fence(struct drm_i915_fence_reg *fence);
> > -void i915_gem_restore_fences(struct drm_i915_private *dev_priv);
> > +void i915_gem_restore_fences(struct drm_i915_private *i915);
> > -void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *dev_priv);
> > +void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *i915);
> >   void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj,
> >   				       struct sg_table *pages);
> >   void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj,
> > @@ -2715,40 +2715,40 @@ int __must_check i915_gem_evict_for_node(struct i915_address_space *vm,
> >   					 unsigned int flags);
> >   int i915_gem_evict_vm(struct i915_address_space *vm);
> > -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv);
> > +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915);
> >   /* belongs in i915_gem_gtt.h */
> > -static inline void i915_gem_chipset_flush(struct drm_i915_private *dev_priv)
> > +static inline void i915_gem_chipset_flush(struct drm_i915_private *i915)
> >   {
> >   	wmb();
> > -	if (INTEL_GEN(dev_priv) < 6)
> > +	if (INTEL_GEN(i915) < 6)
> >   		intel_gtt_chipset_flush();
> >   }
> >   /* i915_gem_stolen.c */
> > -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv,
> > +int i915_gem_stolen_insert_node(struct drm_i915_private *i915,
> >   				struct drm_mm_node *node, u64 size,
> >   				unsigned alignment);
> > -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv,
> > +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915,
> >   					 struct drm_mm_node *node, u64 size,
> >   					 unsigned alignment, u64 start,
> >   					 u64 end);
> > -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv,
> > +void i915_gem_stolen_remove_node(struct drm_i915_private *i915,
> >   				 struct drm_mm_node *node);
> > -int i915_gem_init_stolen(struct drm_i915_private *dev_priv);
> > -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv);
> > +int i915_gem_init_stolen(struct drm_i915_private *i915);
> > +void i915_gem_cleanup_stolen(struct drm_i915_private *i915);
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_stolen(struct drm_i915_private *i915,
> >   			      resource_size_t size);
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915,
> >   					       resource_size_t stolen_offset,
> >   					       resource_size_t gtt_offset,
> >   					       resource_size_t size);
> >   /* i915_gem_internal.c */
> >   struct drm_i915_gem_object *
> > -i915_gem_object_create_internal(struct drm_i915_private *dev_priv,
> > +i915_gem_object_create_internal(struct drm_i915_private *i915,
> >   				phys_addr_t size);
> >   /* i915_gem_shrinker.c */
> > @@ -2771,21 +2771,21 @@ void i915_gem_shrinker_taints_mutex(struct drm_i915_private *i915,
> >   /* i915_gem_tiling.c */
> >   static inline bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> > -	return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 &&
> > +	return i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 &&
> >   		i915_gem_object_is_tiled(obj);
> >   }
> > -u32 i915_gem_fence_size(struct drm_i915_private *dev_priv, u32 size,
> > +u32 i915_gem_fence_size(struct drm_i915_private *i915, u32 size,
> >   			unsigned int tiling, unsigned int stride);
> > -u32 i915_gem_fence_alignment(struct drm_i915_private *dev_priv, u32 size,
> > +u32 i915_gem_fence_alignment(struct drm_i915_private *i915, u32 size,
> >   			     unsigned int tiling, unsigned int stride);
> >   const char *i915_cache_level_str(struct drm_i915_private *i915, int type);
> >   /* i915_cmd_parser.c */
> > -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv);
> > +int i915_cmd_parser_get_version(struct drm_i915_private *i915);
> >   void intel_engine_init_cmd_parser(struct intel_engine_cs *engine);
> >   void intel_engine_cleanup_cmd_parser(struct intel_engine_cs *engine);
> >   int intel_engine_cmd_parser(struct intel_engine_cs *engine,
> > @@ -2796,52 +2796,52 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
> >   			    bool is_master);
> >   /* i915_perf.c */
> > -extern void i915_perf_init(struct drm_i915_private *dev_priv);
> > -extern void i915_perf_fini(struct drm_i915_private *dev_priv);
> > -extern void i915_perf_register(struct drm_i915_private *dev_priv);
> > -extern void i915_perf_unregister(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_init(struct drm_i915_private *i915);
> > +extern void i915_perf_fini(struct drm_i915_private *i915);
> > +extern void i915_perf_register(struct drm_i915_private *i915);
> > +extern void i915_perf_unregister(struct drm_i915_private *i915);
> >   /* i915_suspend.c */
> > -extern int i915_save_state(struct drm_i915_private *dev_priv);
> > -extern int i915_restore_state(struct drm_i915_private *dev_priv);
> > +extern int i915_save_state(struct drm_i915_private *i915);
> > +extern int i915_restore_state(struct drm_i915_private *i915);
> >   /* i915_sysfs.c */
> > -void i915_setup_sysfs(struct drm_i915_private *dev_priv);
> > -void i915_teardown_sysfs(struct drm_i915_private *dev_priv);
> > +void i915_setup_sysfs(struct drm_i915_private *i915);
> > +void i915_teardown_sysfs(struct drm_i915_private *i915);
> >   /* intel_device_info.c */
> >   static inline struct intel_device_info *
> > -mkwrite_device_info(struct drm_i915_private *dev_priv)
> > +mkwrite_device_info(struct drm_i915_private *i915)
> >   {
> > -	return (struct intel_device_info *)INTEL_INFO(dev_priv);
> > +	return (struct intel_device_info *)INTEL_INFO(i915);
> >   }
> >   /* modesetting */
> >   extern void intel_modeset_init_hw(struct drm_device *dev);
> >   extern int intel_modeset_init(struct drm_device *dev);
> >   extern void intel_modeset_cleanup(struct drm_device *dev);
> > -extern int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv,
> > +extern int intel_modeset_vga_set_state(struct drm_i915_private *i915,
> >   				       bool state);
> >   extern void intel_display_resume(struct drm_device *dev);
> > -extern void i915_redisable_vga(struct drm_i915_private *dev_priv);
> > -extern void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv);
> > -extern void intel_init_pch_refclk(struct drm_i915_private *dev_priv);
> > +extern void i915_redisable_vga(struct drm_i915_private *i915);
> > +extern void i915_redisable_vga_power_on(struct drm_i915_private *i915);
> > +extern void intel_init_pch_refclk(struct drm_i915_private *i915);
> >   int i915_reg_read_ioctl(struct drm_device *dev, void *data,
> >   			struct drm_file *file);
> >   extern struct intel_display_error_state *
> > -intel_display_capture_error_state(struct drm_i915_private *dev_priv);
> > +intel_display_capture_error_state(struct drm_i915_private *i915);
> >   extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> >   					    struct intel_display_error_state *error);
> >   #define __I915_REG_OP(op__, dev_priv__, ...) \
> >   	intel_uncore_##op__(&(dev_priv__)->uncore, __VA_ARGS__)
> > -#define I915_READ(reg__)	 __I915_REG_OP(read, dev_priv, (reg__))
> > -#define I915_WRITE(reg__, val__) __I915_REG_OP(write, dev_priv, (reg__), (val__))
> > +#define I915_READ(reg__)	 __I915_REG_OP(read, i915, (reg__))
> > +#define I915_WRITE(reg__, val__) __I915_REG_OP(write, i915, (reg__), (val__))
> > -#define POSTING_READ(reg__)	__I915_REG_OP(posting_read, dev_priv, (reg__))
> > +#define POSTING_READ(reg__)	__I915_REG_OP(posting_read, i915, (reg__))
> >   /* These are untraced mmio-accessors that are only valid to be used inside
> >    * critical sections, such as inside IRQ handlers, where forcewake is explicitly
> > @@ -2851,13 +2851,13 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> >    *
> >    * As an example, these accessors can possibly be used between:
> >    *
> > - * spin_lock_irq(&dev_priv->uncore.lock);
> > + * spin_lock_irq(&i915->uncore.lock);
> >    * intel_uncore_forcewake_get__locked();
> >    *
> >    * and
> >    *
> >    * intel_uncore_forcewake_put__locked();
> > - * spin_unlock_irq(&dev_priv->uncore.lock);
> > + * spin_unlock_irq(&i915->uncore.lock);
> >    *
> >    *
> >    * Note: some registers may not need forcewake held, so
> > @@ -2866,18 +2866,18 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e,
> >    *
> >    * Certain architectures will die if the same cacheline is concurrently accessed
> >    * by different clients (e.g. on Ivybridge). Access to registers should
> > - * therefore generally be serialised, by either the dev_priv->uncore.lock or
> > + * therefore generally be serialised, by either the i915->uncore.lock or
> >    * a more localised lock guarding all access to that bank of registers.
> >    */
> > -#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, dev_priv, (reg__))
> > -#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, dev_priv, (reg__), (val__))
> > +#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, i915, (reg__))
> > +#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, i915, (reg__), (val__))
> >   /* "Broadcast RGB" property */
> >   #define INTEL_BROADCAST_RGB_AUTO 0
> >   #define INTEL_BROADCAST_RGB_FULL 1
> >   #define INTEL_BROADCAST_RGB_LIMITED 2
> > -void i915_memcpy_init_early(struct drm_i915_private *dev_priv);
> > +void i915_memcpy_init_early(struct drm_i915_private *i915);
> >   bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len);
> >   /* The movntdqa instructions used for memcpy-from-wc require 16-byte alignment,
> > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> > index 4017ecf561f6..1774c7604076 100644
> > --- a/drivers/gpu/drm/i915/i915_gem.c
> > +++ b/drivers/gpu/drm/i915/i915_gem.c
> > @@ -149,7 +149,7 @@ i915_gem_phys_pwrite(struct drm_i915_gem_object *obj,
> >   static int
> >   i915_gem_create(struct drm_file *file,
> > -		struct drm_i915_private *dev_priv,
> > +		struct drm_i915_private *i915,
> >   		u64 *size_p,
> >   		u32 *handle_p)
> >   {
> > @@ -163,7 +163,7 @@ i915_gem_create(struct drm_file *file,
> >   		return -EINVAL;
> >   	/* Allocate the new object */
> > -	obj = i915_gem_object_create_shmem(dev_priv, size);
> > +	obj = i915_gem_object_create_shmem(i915, size);
> >   	if (IS_ERR(obj))
> >   		return PTR_ERR(obj);
> > @@ -223,16 +223,16 @@ int
> >   i915_gem_create_ioctl(struct drm_device *dev, void *data,
> >   		      struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_create *args = data;
> > -	i915_gem_flush_free_objects(dev_priv);
> > +	i915_gem_flush_free_objects(i915);
> > -	return i915_gem_create(file, dev_priv,
> > +	return i915_gem_create(file, i915,
> >   			       &args->size, &args->handle);
> >   }
> > -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv)
> > +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref;
> > @@ -257,13 +257,13 @@ void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv)
> >   	wmb();
> > -	if (INTEL_INFO(dev_priv)->has_coherent_ggtt)
> > +	if (INTEL_INFO(i915)->has_coherent_ggtt)
> >   		return;
> > -	i915_gem_chipset_flush(dev_priv);
> > +	i915_gem_chipset_flush(i915);
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > -		struct intel_uncore *uncore = &dev_priv->uncore;
> > +	with_intel_runtime_pm(i915, wakeref) {
> > +		struct intel_uncore *uncore = &i915->uncore;
> >   		spin_lock_irq(&uncore->lock);
> >   		intel_uncore_posting_read_fw(uncore,
> > @@ -884,7 +884,7 @@ i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data,
> >   	return 0;
> >   }
> > -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv)
> > +void i915_gem_runtime_suspend(struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_gem_object *obj, *on;
> >   	int i;
> > @@ -897,15 +897,15 @@ void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv)
> >   	 */
> >   	list_for_each_entry_safe(obj, on,
> > -				 &dev_priv->mm.userfault_list, userfault_link)
> > +				 &i915->mm.userfault_list, userfault_link)
> >   		__i915_gem_object_release_mmap(obj);
> >   	/* The fence will be lost when the device powers down. If any were
> >   	 * in use by hardware (i.e. they are pinned), we should not be powering
> >   	 * down! All other fences will be reacquired by the user upon waking.
> >   	 */
> > -	for (i = 0; i < dev_priv->num_fence_regs; i++) {
> > -		struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i];
> > +	for (i = 0; i < i915->num_fence_regs; i++) {
> > +		struct drm_i915_fence_reg *reg = &i915->fence_regs[i];
> >   		/* Ideally we want to assert that the fence register is not
> >   		 * live at this point (i.e. that no piece of code will be
> > @@ -1020,8 +1020,8 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> >   			 u64 alignment,
> >   			 u64 flags)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > -	struct i915_address_space *vm = &dev_priv->ggtt.vm;
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> > +	struct i915_address_space *vm = &i915->ggtt.vm;
> >   	struct i915_vma *vma;
> >   	int ret;
> > @@ -1036,7 +1036,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> >   		 * the object in and out of the Global GTT and
> >   		 * waste a lot of cycles under the mutex.
> >   		 */
> > -		if (obj->base.size > dev_priv->ggtt.mappable_end)
> > +		if (obj->base.size > i915->ggtt.mappable_end)
> >   			return ERR_PTR(-E2BIG);
> >   		/* If NONBLOCK is set the caller is optimistically
> > @@ -1055,7 +1055,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> >   		 * we could try to minimise harm to others.
> >   		 */
> >   		if (flags & PIN_NONBLOCK &&
> > -		    obj->base.size > dev_priv->ggtt.mappable_end / 2)
> > +		    obj->base.size > i915->ggtt.mappable_end / 2)
> >   			return ERR_PTR(-ENOSPC);
> >   	}
> > @@ -1069,7 +1069,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj,
> >   				return ERR_PTR(-ENOSPC);
> >   			if (flags & PIN_MAPPABLE &&
> > -			    vma->fence_size > dev_priv->ggtt.mappable_end / 2)
> > +			    vma->fence_size > i915->ggtt.mappable_end / 2)
> >   				return ERR_PTR(-ENOSPC);
> >   		}
> > @@ -1202,30 +1202,30 @@ void i915_gem_sanitize(struct drm_i915_private *i915)
> >   	mutex_unlock(&i915->drm.struct_mutex);
> >   }
> > -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv)
> > +void i915_gem_init_swizzling(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) < 5 ||
> > -	    dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE)
> > +	if (INTEL_GEN(i915) < 5 ||
> > +	    i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE)
> >   		return;
> >   	I915_WRITE(DISP_ARB_CTL, I915_READ(DISP_ARB_CTL) |
> >   				 DISP_TILE_SURFACE_SWIZZLING);
> > -	if (IS_GEN(dev_priv, 5))
> > +	if (IS_GEN(i915, 5))
> >   		return;
> >   	I915_WRITE(TILECTL, I915_READ(TILECTL) | TILECTL_SWZCTL);
> > -	if (IS_GEN(dev_priv, 6))
> > +	if (IS_GEN(i915, 6))
> >   		I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_SNB));
> > -	else if (IS_GEN(dev_priv, 7))
> > +	else if (IS_GEN(i915, 7))
> >   		I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_IVB));
> > -	else if (IS_GEN(dev_priv, 8))
> > +	else if (IS_GEN(i915, 8))
> >   		I915_WRITE(GAMTARBMODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_BDW));
> >   	else
> >   		BUG();
> >   }
> > -static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base)
> > +static void init_unused_ring(struct drm_i915_private *i915, u32 base)
> >   {
> >   	I915_WRITE(RING_CTL(base), 0);
> >   	I915_WRITE(RING_HEAD(base), 0);
> > @@ -1233,45 +1233,45 @@ static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base)
> >   	I915_WRITE(RING_START(base), 0);
> >   }
> > -static void init_unused_rings(struct drm_i915_private *dev_priv)
> > +static void init_unused_rings(struct drm_i915_private *i915)
> >   {
> > -	if (IS_I830(dev_priv)) {
> > -		init_unused_ring(dev_priv, PRB1_BASE);
> > -		init_unused_ring(dev_priv, SRB0_BASE);
> > -		init_unused_ring(dev_priv, SRB1_BASE);
> > -		init_unused_ring(dev_priv, SRB2_BASE);
> > -		init_unused_ring(dev_priv, SRB3_BASE);
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > -		init_unused_ring(dev_priv, SRB0_BASE);
> > -		init_unused_ring(dev_priv, SRB1_BASE);
> > -	} else if (IS_GEN(dev_priv, 3)) {
> > -		init_unused_ring(dev_priv, PRB1_BASE);
> > -		init_unused_ring(dev_priv, PRB2_BASE);
> > +	if (IS_I830(i915)) {
> > +		init_unused_ring(i915, PRB1_BASE);
> > +		init_unused_ring(i915, SRB0_BASE);
> > +		init_unused_ring(i915, SRB1_BASE);
> > +		init_unused_ring(i915, SRB2_BASE);
> > +		init_unused_ring(i915, SRB3_BASE);
> > +	} else if (IS_GEN(i915, 2)) {
> > +		init_unused_ring(i915, SRB0_BASE);
> > +		init_unused_ring(i915, SRB1_BASE);
> > +	} else if (IS_GEN(i915, 3)) {
> > +		init_unused_ring(i915, PRB1_BASE);
> > +		init_unused_ring(i915, PRB2_BASE);
> >   	}
> >   }
> > -int i915_gem_init_hw(struct drm_i915_private *dev_priv)
> > +int i915_gem_init_hw(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	dev_priv->gt.last_init_time = ktime_get();
> > +	i915->gt.last_init_time = ktime_get();
> >   	/* Double layer security blanket, see i915_gem_init() */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> > -	if (HAS_EDRAM(dev_priv) && INTEL_GEN(dev_priv) < 9)
> > +	if (HAS_EDRAM(i915) && INTEL_GEN(i915) < 9)
> >   		I915_WRITE(HSW_IDICR, I915_READ(HSW_IDICR) | IDIHASHMSK(0xf));
> > -	if (IS_HASWELL(dev_priv))
> > -		I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(dev_priv) ?
> > +	if (IS_HASWELL(i915))
> > +		I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(i915) ?
> >   			   LOWER_SLICE_ENABLED : LOWER_SLICE_DISABLED);
> >   	/* Apply the GT workarounds... */
> > -	intel_gt_apply_workarounds(dev_priv);
> > +	intel_gt_apply_workarounds(i915);
> >   	/* ...and determine whether they are sticking. */
> > -	intel_gt_verify_workarounds(dev_priv, "init");
> > +	intel_gt_verify_workarounds(i915, "init");
> > -	i915_gem_init_swizzling(dev_priv);
> > +	i915_gem_init_swizzling(i915);
> >   	/*
> >   	 * At least 830 can leave some of the unused rings
> > @@ -1279,48 +1279,48 @@ int i915_gem_init_hw(struct drm_i915_private *dev_priv)
> >   	 * will prevent c3 entry. Makes sure all unused rings
> >   	 * are totally idle.
> >   	 */
> > -	init_unused_rings(dev_priv);
> > +	init_unused_rings(i915);
> > -	BUG_ON(!dev_priv->kernel_context);
> > -	ret = i915_terminally_wedged(dev_priv);
> > +	BUG_ON(!i915->kernel_context);
> > +	ret = i915_terminally_wedged(i915);
> >   	if (ret)
> >   		goto out;
> > -	ret = i915_ppgtt_init_hw(dev_priv);
> > +	ret = i915_ppgtt_init_hw(i915);
> >   	if (ret) {
> >   		DRM_ERROR("Enabling PPGTT failed (%d)\n", ret);
> >   		goto out;
> >   	}
> > -	ret = intel_wopcm_init_hw(&dev_priv->wopcm);
> > +	ret = intel_wopcm_init_hw(&i915->wopcm);
> >   	if (ret) {
> >   		DRM_ERROR("Enabling WOPCM failed (%d)\n", ret);
> >   		goto out;
> >   	}
> >   	/* We can't enable contexts until all firmware is loaded */
> > -	ret = intel_uc_init_hw(dev_priv);
> > +	ret = intel_uc_init_hw(i915);
> >   	if (ret) {
> >   		DRM_ERROR("Enabling uc failed (%d)\n", ret);
> >   		goto out;
> >   	}
> > -	intel_mocs_init_l3cc_table(dev_priv);
> > +	intel_mocs_init_l3cc_table(i915);
> >   	/* Only when the HW is re-initialised, can we replay the requests */
> > -	ret = intel_engines_resume(dev_priv);
> > +	ret = intel_engines_resume(i915);
> >   	if (ret)
> >   		goto cleanup_uc;
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > -	intel_engines_set_scheduler_caps(dev_priv);
> > +	intel_engines_set_scheduler_caps(i915);
> >   	return 0;
> >   cleanup_uc:
> > -	intel_uc_fini_hw(dev_priv);
> > +	intel_uc_fini_hw(i915);
> >   out:
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   	return ret;
> >   }
> > @@ -1505,28 +1505,28 @@ static int intel_engines_verify_workarounds(struct drm_i915_private *i915)
> >   	return err;
> >   }
> > -int i915_gem_init(struct drm_i915_private *dev_priv)
> > +int i915_gem_init(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> >   	/* We need to fallback to 4K pages if host doesn't support huge gtt. */
> > -	if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv))
> > -		mkwrite_device_info(dev_priv)->page_sizes =
> > +	if (intel_vgpu_active(i915) && !intel_vgpu_has_huge_gtt(i915))
> > +		mkwrite_device_info(i915)->page_sizes =
> >   			I915_GTT_PAGE_SIZE_4K;
> > -	dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1);
> > +	i915->mm.unordered_timeline = dma_fence_context_alloc(1);
> > -	i915_timelines_init(dev_priv);
> > +	i915_timelines_init(i915);
> > -	ret = i915_gem_init_userptr(dev_priv);
> > +	ret = i915_gem_init_userptr(i915);
> >   	if (ret)
> >   		return ret;
> > -	ret = intel_uc_init_misc(dev_priv);
> > +	ret = intel_uc_init_misc(i915);
> >   	if (ret)
> >   		return ret;
> > -	ret = intel_wopcm_init(&dev_priv->wopcm);
> > +	ret = intel_wopcm_init(&i915->wopcm);
> >   	if (ret)
> >   		goto err_uc_misc;
> > @@ -1536,47 +1536,47 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> >   	 * we hold the forcewake during initialisation these problems
> >   	 * just magically go away.
> >   	 */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> > -	ret = i915_gem_init_ggtt(dev_priv);
> > +	ret = i915_gem_init_ggtt(i915);
> >   	if (ret) {
> >   		GEM_BUG_ON(ret == -EIO);
> >   		goto err_unlock;
> >   	}
> > -	ret = i915_gem_init_scratch(dev_priv,
> > -				    IS_GEN(dev_priv, 2) ? SZ_256K : PAGE_SIZE);
> > +	ret = i915_gem_init_scratch(i915,
> > +				    IS_GEN(i915, 2) ? SZ_256K : PAGE_SIZE);
> >   	if (ret) {
> >   		GEM_BUG_ON(ret == -EIO);
> >   		goto err_ggtt;
> >   	}
> > -	ret = intel_engines_setup(dev_priv);
> > +	ret = intel_engines_setup(i915);
> >   	if (ret) {
> >   		GEM_BUG_ON(ret == -EIO);
> >   		goto err_unlock;
> >   	}
> > -	ret = i915_gem_contexts_init(dev_priv);
> > +	ret = i915_gem_contexts_init(i915);
> >   	if (ret) {
> >   		GEM_BUG_ON(ret == -EIO);
> >   		goto err_scratch;
> >   	}
> > -	ret = intel_engines_init(dev_priv);
> > +	ret = intel_engines_init(i915);
> >   	if (ret) {
> >   		GEM_BUG_ON(ret == -EIO);
> >   		goto err_context;
> >   	}
> > -	intel_init_gt_powersave(dev_priv);
> > +	intel_init_gt_powersave(i915);
> > -	ret = intel_uc_init(dev_priv);
> > +	ret = intel_uc_init(i915);
> >   	if (ret)
> >   		goto err_pm;
> > -	ret = i915_gem_init_hw(dev_priv);
> > +	ret = i915_gem_init_hw(i915);
> >   	if (ret)
> >   		goto err_uc_init;
> > @@ -1589,13 +1589,13 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> >   	 *
> >   	 * FIXME: break up the workarounds and apply them at the right time!
> >   	 */
> > -	intel_init_clock_gating(dev_priv);
> > +	intel_init_clock_gating(i915);
> > -	ret = intel_engines_verify_workarounds(dev_priv);
> > +	ret = intel_engines_verify_workarounds(i915);
> >   	if (ret)
> >   		goto err_init_hw;
> > -	ret = __intel_engines_record_defaults(dev_priv);
> > +	ret = __intel_engines_record_defaults(i915);
> >   	if (ret)
> >   		goto err_init_hw;
> > @@ -1609,8 +1609,8 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> >   		goto err_init_hw;
> >   	}
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return 0;
> > @@ -1621,107 +1621,107 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> >   	 * driver doesn't explode during runtime.
> >   	 */
> >   err_init_hw:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	i915_gem_set_wedged(dev_priv);
> > -	i915_gem_suspend(dev_priv);
> > -	i915_gem_suspend_late(dev_priv);
> > +	i915_gem_set_wedged(i915);
> > +	i915_gem_suspend(i915);
> > +	i915_gem_suspend_late(i915);
> > -	i915_gem_drain_workqueue(dev_priv);
> > +	i915_gem_drain_workqueue(i915);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	intel_uc_fini_hw(dev_priv);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	intel_uc_fini_hw(i915);
> >   err_uc_init:
> > -	intel_uc_fini(dev_priv);
> > +	intel_uc_fini(i915);
> >   err_pm:
> >   	if (ret != -EIO) {
> > -		intel_cleanup_gt_powersave(dev_priv);
> > -		intel_engines_cleanup(dev_priv);
> > +		intel_cleanup_gt_powersave(i915);
> > +		intel_engines_cleanup(i915);
> >   	}
> >   err_context:
> >   	if (ret != -EIO)
> > -		i915_gem_contexts_fini(dev_priv);
> > +		i915_gem_contexts_fini(i915);
> >   err_scratch:
> > -	i915_gem_fini_scratch(dev_priv);
> > +	i915_gem_fini_scratch(i915);
> >   err_ggtt:
> >   err_unlock:
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   err_uc_misc:
> > -	intel_uc_fini_misc(dev_priv);
> > +	intel_uc_fini_misc(i915);
> >   	if (ret != -EIO) {
> > -		i915_gem_cleanup_userptr(dev_priv);
> > -		i915_timelines_fini(dev_priv);
> > +		i915_gem_cleanup_userptr(i915);
> > +		i915_timelines_fini(i915);
> >   	}
> >   	if (ret == -EIO) {
> > -		mutex_lock(&dev_priv->drm.struct_mutex);
> > +		mutex_lock(&i915->drm.struct_mutex);
> >   		/*
> >   		 * Allow engine initialisation to fail by marking the GPU as
> >   		 * wedged. But we only want to do this where the GPU is angry,
> >   		 * for all other failure, such as an allocation failure, bail.
> >   		 */
> > -		if (!i915_reset_failed(dev_priv)) {
> > -			i915_load_error(dev_priv,
> > +		if (!i915_reset_failed(i915)) {
> > +			i915_load_error(i915,
> >   					"Failed to initialize GPU, declaring it wedged!\n");
> > -			i915_gem_set_wedged(dev_priv);
> > +			i915_gem_set_wedged(i915);
> >   		}
> >   		/* Minimal basic recovery for KMS */
> > -		ret = i915_ggtt_enable_hw(dev_priv);
> > -		i915_gem_restore_gtt_mappings(dev_priv);
> > -		i915_gem_restore_fences(dev_priv);
> > -		intel_init_clock_gating(dev_priv);
> > +		ret = i915_ggtt_enable_hw(i915);
> > +		i915_gem_restore_gtt_mappings(i915);
> > +		i915_gem_restore_fences(i915);
> > +		intel_init_clock_gating(i915);
> > -		mutex_unlock(&dev_priv->drm.struct_mutex);
> > +		mutex_unlock(&i915->drm.struct_mutex);
> >   	}
> > -	i915_gem_drain_freed_objects(dev_priv);
> > +	i915_gem_drain_freed_objects(i915);
> >   	return ret;
> >   }
> > -void i915_gem_fini_hw(struct drm_i915_private *dev_priv)
> > +void i915_gem_fini_hw(struct drm_i915_private *i915)
> >   {
> > -	GEM_BUG_ON(dev_priv->gt.awake);
> > +	GEM_BUG_ON(i915->gt.awake);
> > -	intel_wakeref_auto_fini(&dev_priv->mm.userfault_wakeref);
> > +	intel_wakeref_auto_fini(&i915->mm.userfault_wakeref);
> > -	i915_gem_suspend_late(dev_priv);
> > -	intel_disable_gt_powersave(dev_priv);
> > +	i915_gem_suspend_late(i915);
> > +	intel_disable_gt_powersave(i915);
> >   	/* Flush any outstanding unpin_work. */
> > -	i915_gem_drain_workqueue(dev_priv);
> > +	i915_gem_drain_workqueue(i915);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	intel_uc_fini_hw(dev_priv);
> > -	intel_uc_fini(dev_priv);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	intel_uc_fini_hw(i915);
> > +	intel_uc_fini(i915);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	i915_gem_drain_freed_objects(dev_priv);
> > +	i915_gem_drain_freed_objects(i915);
> >   }
> > -void i915_gem_fini(struct drm_i915_private *dev_priv)
> > +void i915_gem_fini(struct drm_i915_private *i915)
> >   {
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	intel_engines_cleanup(dev_priv);
> > -	i915_gem_contexts_fini(dev_priv);
> > -	i915_gem_fini_scratch(dev_priv);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	intel_engines_cleanup(i915);
> > +	i915_gem_contexts_fini(i915);
> > +	i915_gem_fini_scratch(i915);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	intel_wa_list_free(&dev_priv->gt_wa_list);
> > +	intel_wa_list_free(&i915->gt_wa_list);
> > -	intel_cleanup_gt_powersave(dev_priv);
> > +	intel_cleanup_gt_powersave(i915);
> > -	intel_uc_fini_misc(dev_priv);
> > -	i915_gem_cleanup_userptr(dev_priv);
> > -	i915_timelines_fini(dev_priv);
> > +	intel_uc_fini_misc(i915);
> > +	i915_gem_cleanup_userptr(i915);
> > +	i915_timelines_fini(i915);
> > -	i915_gem_drain_freed_objects(dev_priv);
> > +	i915_gem_drain_freed_objects(i915);
> > -	WARN_ON(!list_empty(&dev_priv->contexts.list));
> > +	WARN_ON(!list_empty(&i915->contexts.list));
> >   }
> >   void i915_gem_init_mmio(struct drm_i915_private *i915)
> > @@ -1730,35 +1730,35 @@ void i915_gem_init_mmio(struct drm_i915_private *i915)
> >   }
> >   void
> > -i915_gem_load_init_fences(struct drm_i915_private *dev_priv)
> > +i915_gem_load_init_fences(struct drm_i915_private *i915)
> >   {
> >   	int i;
> > -	if (INTEL_GEN(dev_priv) >= 7 && !IS_VALLEYVIEW(dev_priv) &&
> > -	    !IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->num_fence_regs = 32;
> > -	else if (INTEL_GEN(dev_priv) >= 4 ||
> > -		 IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> > -		 IS_G33(dev_priv) || IS_PINEVIEW(dev_priv))
> > -		dev_priv->num_fence_regs = 16;
> > +	if (INTEL_GEN(i915) >= 7 && !IS_VALLEYVIEW(i915) &&
> > +	    !IS_CHERRYVIEW(i915))
> > +		i915->num_fence_regs = 32;
> > +	else if (INTEL_GEN(i915) >= 4 ||
> > +		 IS_I945G(i915) || IS_I945GM(i915) ||
> > +		 IS_G33(i915) || IS_PINEVIEW(i915))
> > +		i915->num_fence_regs = 16;
> >   	else
> > -		dev_priv->num_fence_regs = 8;
> > +		i915->num_fence_regs = 8;
> > -	if (intel_vgpu_active(dev_priv))
> > -		dev_priv->num_fence_regs =
> > +	if (intel_vgpu_active(i915))
> > +		i915->num_fence_regs =
> >   				I915_READ(vgtif_reg(avail_rs.fence_num));
> >   	/* Initialize fence registers to zero */
> > -	for (i = 0; i < dev_priv->num_fence_regs; i++) {
> > -		struct drm_i915_fence_reg *fence = &dev_priv->fence_regs[i];
> > +	for (i = 0; i < i915->num_fence_regs; i++) {
> > +		struct drm_i915_fence_reg *fence = &i915->fence_regs[i];
> > -		fence->i915 = dev_priv;
> > +		fence->i915 = i915;
> >   		fence->id = i;
> > -		list_add_tail(&fence->link, &dev_priv->mm.fence_list);
> > +		list_add_tail(&fence->link, &i915->mm.fence_list);
> >   	}
> > -	i915_gem_restore_fences(dev_priv);
> > +	i915_gem_restore_fences(i915);
> > -	i915_gem_detect_bit_6_swizzle(dev_priv);
> > +	i915_gem_detect_bit_6_swizzle(i915);
> >   }
> >   static void i915_gem_init__mm(struct drm_i915_private *i915)
> > @@ -1778,56 +1778,56 @@ static void i915_gem_init__mm(struct drm_i915_private *i915)
> >   	i915_gem_init__objects(i915);
> >   }
> > -int i915_gem_init_early(struct drm_i915_private *dev_priv)
> > +int i915_gem_init_early(struct drm_i915_private *i915)
> >   {
> >   	static struct lock_class_key reset_key;
> >   	int err;
> > -	intel_gt_pm_init(dev_priv);
> > +	intel_gt_pm_init(i915);
> > -	INIT_LIST_HEAD(&dev_priv->gt.active_rings);
> > -	INIT_LIST_HEAD(&dev_priv->gt.closed_vma);
> > -	spin_lock_init(&dev_priv->gt.closed_lock);
> > -	lockdep_init_map(&dev_priv->gt.reset_lockmap,
> > +	INIT_LIST_HEAD(&i915->gt.active_rings);
> > +	INIT_LIST_HEAD(&i915->gt.closed_vma);
> > +	spin_lock_init(&i915->gt.closed_lock);
> > +	lockdep_init_map(&i915->gt.reset_lockmap,
> >   			 "i915.reset", &reset_key, 0);
> > -	i915_gem_init__mm(dev_priv);
> > -	i915_gem_init__pm(dev_priv);
> > +	i915_gem_init__mm(i915);
> > +	i915_gem_init__pm(i915);
> > -	init_waitqueue_head(&dev_priv->gpu_error.wait_queue);
> > -	init_waitqueue_head(&dev_priv->gpu_error.reset_queue);
> > -	mutex_init(&dev_priv->gpu_error.wedge_mutex);
> > -	init_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu);
> > +	init_waitqueue_head(&i915->gpu_error.wait_queue);
> > +	init_waitqueue_head(&i915->gpu_error.reset_queue);
> > +	mutex_init(&i915->gpu_error.wedge_mutex);
> > +	init_srcu_struct(&i915->gpu_error.reset_backoff_srcu);
> > -	atomic_set(&dev_priv->mm.bsd_engine_dispatch_index, 0);
> > +	atomic_set(&i915->mm.bsd_engine_dispatch_index, 0);
> > -	spin_lock_init(&dev_priv->fb_tracking.lock);
> > +	spin_lock_init(&i915->fb_tracking.lock);
> > -	err = i915_gemfs_init(dev_priv);
> > +	err = i915_gemfs_init(i915);
> >   	if (err)
> >   		DRM_NOTE("Unable to create a private tmpfs mount, hugepage support will be disabled(%d).\n", err);
> >   	return 0;
> >   }
> > -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv)
> > +void i915_gem_cleanup_early(struct drm_i915_private *i915)
> >   {
> > -	i915_gem_drain_freed_objects(dev_priv);
> > -	GEM_BUG_ON(!llist_empty(&dev_priv->mm.free_list));
> > -	GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count));
> > -	WARN_ON(dev_priv->mm.shrink_count);
> > +	i915_gem_drain_freed_objects(i915);
> > +	GEM_BUG_ON(!llist_empty(&i915->mm.free_list));
> > +	GEM_BUG_ON(atomic_read(&i915->mm.free_count));
> > +	WARN_ON(i915->mm.shrink_count);
> > -	cleanup_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu);
> > +	cleanup_srcu_struct(&i915->gpu_error.reset_backoff_srcu);
> > -	i915_gemfs_fini(dev_priv);
> > +	i915_gemfs_fini(i915);
> >   }
> > -int i915_gem_freeze(struct drm_i915_private *dev_priv)
> > +int i915_gem_freeze(struct drm_i915_private *i915)
> >   {
> >   	/* Discard all purgeable objects, let userspace recover those as
> >   	 * required after resuming.
> >   	 */
> > -	i915_gem_shrink_all(dev_priv);
> > +	i915_gem_shrink_all(i915);
> >   	return 0;
> >   }
> > @@ -1895,7 +1895,7 @@ int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file)
> >   		return -ENOMEM;
> >   	file->driver_priv = file_priv;
> > -	file_priv->dev_priv = i915;
> > +	file_priv->i915 = i915;
> >   	file_priv->file = file;
> >   	spin_lock_init(&file_priv->mm.lock);
> > diff --git a/drivers/gpu/drm/i915/i915_gem_evict.c b/drivers/gpu/drm/i915/i915_gem_evict.c
> > index a5783c4cb98b..4244d9c54d66 100644
> > --- a/drivers/gpu/drm/i915/i915_gem_evict.c
> > +++ b/drivers/gpu/drm/i915/i915_gem_evict.c
> > @@ -99,7 +99,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> >   			 u64 start, u64 end,
> >   			 unsigned flags)
> >   {
> > -	struct drm_i915_private *dev_priv = vm->i915;
> > +	struct drm_i915_private *i915 = vm->i915;
> >   	struct drm_mm_scan scan;
> >   	struct list_head eviction_list;
> >   	struct i915_vma *vma, *next;
> > @@ -138,7 +138,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> >   	 * retiring.
> >   	 */
> >   	if (!(flags & PIN_NONBLOCK))
> > -		i915_retire_requests(dev_priv);
> > +		i915_retire_requests(i915);
> >   search_again:
> >   	active = NULL;
> > @@ -211,7 +211,7 @@ i915_gem_evict_something(struct i915_address_space *vm,
> >   	if (I915_SELFTEST_ONLY(igt_evict_ctl.fail_if_busy))
> >   		return -EBUSY;
> > -	ret = ggtt_flush(dev_priv);
> > +	ret = ggtt_flush(i915);
> >   	if (ret)
> >   		return ret;
> > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
> > index e70675bfb51d..42d88dae9c5d 100644
> > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c
> > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
> > @@ -1294,7 +1294,7 @@ static int gen8_init_scratch(struct i915_address_space *vm)
> >   static int gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create)
> >   {
> >   	struct i915_address_space *vm = &ppgtt->vm;
> > -	struct drm_i915_private *dev_priv = vm->i915;
> > +	struct drm_i915_private *i915 = vm->i915;
> >   	enum vgt_g2v_type msg;
> >   	int i;
> > @@ -1692,7 +1692,7 @@ static inline void gen6_write_pde(const struct gen6_ppgtt *ppgtt,
> >   		  ppgtt->pd_addr + pde);
> >   }
> > -static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> > +static void gen7_ppgtt_enable(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	u32 ecochk, ecobits;
> > @@ -1702,7 +1702,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GAC_ECO_BITS, ecobits | ECOBITS_PPGTT_CACHE64B);
> >   	ecochk = I915_READ(GAM_ECOCHK);
> > -	if (IS_HASWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915)) {
> >   		ecochk |= ECOCHK_PPGTT_WB_HSW;
> >   	} else {
> >   		ecochk |= ECOCHK_PPGTT_LLC_IVB;
> > @@ -1710,7 +1710,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> >   	}
> >   	I915_WRITE(GAM_ECOCHK, ecochk);
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		/* GFX_MODE is per-ring on gen7+ */
> >   		ENGINE_WRITE(engine,
> >   			     RING_MODE_GEN7,
> > @@ -1718,7 +1718,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv)
> > +static void gen6_ppgtt_enable(struct drm_i915_private *i915)
> >   {
> >   	u32 ecochk, gab_ctl, ecobits;
> > @@ -1732,7 +1732,7 @@ static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv)
> >   	ecochk = I915_READ(GAM_ECOCHK);
> >   	I915_WRITE(GAM_ECOCHK, ecochk | ECOCHK_SNB_BIT | ECOCHK_PPGTT_CACHE64B);
> > -	if (HAS_PPGTT(dev_priv)) /* may be disabled for VT-d */
> > +	if (HAS_PPGTT(i915)) /* may be disabled for VT-d */
> >   		I915_WRITE(GFX_MODE, _MASKED_BIT_ENABLE(GFX_PPGTT_ENABLE));
> >   }
> > @@ -2185,20 +2185,20 @@ static struct i915_ppgtt *gen6_ppgtt_create(struct drm_i915_private *i915)
> >   	return ERR_PTR(err);
> >   }
> > -static void gtt_write_workarounds(struct drm_i915_private *dev_priv)
> > +static void gtt_write_workarounds(struct drm_i915_private *i915)
> >   {
> >   	/* This function is for gtt related workarounds. This function is
> >   	 * called on driver load and after a GPU reset, so you can place
> >   	 * workarounds here even if they get overwritten by GPU reset.
> >   	 */
> >   	/* WaIncreaseDefaultTLBEntries:chv,bdw,skl,bxt,kbl,glk,cfl,cnl,icl */
> > -	if (IS_BROADWELL(dev_priv))
> > +	if (IS_BROADWELL(i915))
> >   		I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_BDW);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915))
> >   		I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_CHV);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_BXT);
> > -	else if (INTEL_GEN(dev_priv) >= 9)
> > +	else if (INTEL_GEN(i915) >= 9)
> >   		I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_SKL);
> >   	/*
> > @@ -2212,21 +2212,21 @@ static void gtt_write_workarounds(struct drm_i915_private *dev_priv)
> >   	 * 32K pages, but we don't currently have any support for it in our
> >   	 * driver.
> >   	 */
> > -	if (HAS_PAGE_SIZES(dev_priv, I915_GTT_PAGE_SIZE_64K) &&
> > -	    INTEL_GEN(dev_priv) <= 10)
> > +	if (HAS_PAGE_SIZES(i915, I915_GTT_PAGE_SIZE_64K) &&
> > +	    INTEL_GEN(i915) <= 10)
> >   		I915_WRITE(GEN8_GAMW_ECO_DEV_RW_IA,
> >   			   I915_READ(GEN8_GAMW_ECO_DEV_RW_IA) |
> >   			   GAMW_ECO_ENABLE_64K_IPS_FIELD);
> >   }
> > -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv)
> > +int i915_ppgtt_init_hw(struct drm_i915_private *i915)
> >   {
> > -	gtt_write_workarounds(dev_priv);
> > +	gtt_write_workarounds(i915);
> > -	if (IS_GEN(dev_priv, 6))
> > -		gen6_ppgtt_enable(dev_priv);
> > -	else if (IS_GEN(dev_priv, 7))
> > -		gen7_ppgtt_enable(dev_priv);
> > +	if (IS_GEN(i915, 6))
> > +		gen6_ppgtt_enable(i915);
> > +	else if (IS_GEN(i915, 7))
> > +		gen7_ppgtt_enable(i915);
> >   	return 0;
> >   }
> > @@ -2293,29 +2293,29 @@ void i915_vm_release(struct kref *kref)
> >   /* Certain Gen5 chipsets require require idling the GPU before
> >    * unmapping anything from the GTT when VT-d is enabled.
> >    */
> > -static bool needs_idle_maps(struct drm_i915_private *dev_priv)
> > +static bool needs_idle_maps(struct drm_i915_private *i915)
> >   {
> >   	/* Query intel_iommu to see if we need the workaround. Presumably that
> >   	 * was loaded first.
> >   	 */
> > -	return IS_GEN(dev_priv, 5) && IS_MOBILE(dev_priv) && intel_vtd_active();
> > +	return IS_GEN(i915, 5) && IS_MOBILE(i915) && intel_vtd_active();
> >   }
> > -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv)
> > +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	/* Don't bother messing with faults pre GEN6 as we have little
> >   	 * documentation supporting that it's a good idea.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 6)
> > +	if (INTEL_GEN(i915) < 6)
> >   		return;
> > -	i915_check_and_clear_faults(dev_priv);
> > +	i915_check_and_clear_faults(i915);
> >   	ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
> > -	i915_ggtt_invalidate(dev_priv);
> > +	i915_ggtt_invalidate(i915);
> >   }
> >   int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
> > @@ -2411,7 +2411,7 @@ static void gen6_ggtt_insert_page(struct i915_address_space *vm,
> >    * Binds an object into the global gtt with the specified cache level. The object
> >    * will be accessible to the GPU via commands whose operands reference offsets
> >    * within the global GTT as well as accessible by the GPU through the GMADR
> > - * mapped BAR (dev_priv->mm.gtt->gtt).
> > + * mapped BAR (i915->mm.gtt->gtt).
> >    */
> >   static void gen6_ggtt_insert_entries(struct i915_address_space *vm,
> >   				     struct i915_vma *vma,
> > @@ -2461,7 +2461,7 @@ static void gen8_ggtt_clear_range(struct i915_address_space *vm,
> >   static void bxt_vtd_ggtt_wa(struct i915_address_space *vm)
> >   {
> > -	struct drm_i915_private *dev_priv = vm->i915;
> > +	struct drm_i915_private *i915 = vm->i915;
> >   	/*
> >   	 * Make sure the internal GAM fifo has been cleared of all GTT
> > @@ -2705,12 +2705,12 @@ static void aliasing_gtt_unbind_vma(struct i915_vma *vma)
> >   void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj,
> >   			       struct sg_table *pages)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > -	struct device *kdev = &dev_priv->drm.pdev->dev;
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> > +	struct device *kdev = &i915->drm.pdev->dev;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	if (unlikely(ggtt->do_idle_maps)) {
> > -		if (i915_gem_wait_for_idle(dev_priv, 0, MAX_SCHEDULE_TIMEOUT)) {
> > +		if (i915_gem_wait_for_idle(i915, 0, MAX_SCHEDULE_TIMEOUT)) {
> >   			DRM_ERROR("Failed to wait for idle; VT'd may hang.\n");
> >   			/* Wait a bit, in hopes it avoids the hang */
> >   			udelay(10);
> > @@ -2834,7 +2834,7 @@ static void ggtt_release_guc_top(struct i915_ggtt *ggtt)
> >   		drm_mm_remove_node(&ggtt->uc_fw);
> >   }
> > -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> > +int i915_gem_init_ggtt(struct drm_i915_private *i915)
> >   {
> >   	/* Let GEM Manage all of the aperture.
> >   	 *
> > @@ -2845,7 +2845,7 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> >   	 * aperture.  One page should be enough to keep any prefetching inside
> >   	 * of the aperture.
> >   	 */
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	unsigned long hole_start, hole_end;
> >   	struct drm_mm_node *entry;
> >   	int ret;
> > @@ -2857,9 +2857,9 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> >   	 * why.
> >   	 */
> >   	ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE,
> > -			       intel_wopcm_guc_size(&dev_priv->wopcm));
> > +			       intel_wopcm_guc_size(&i915->wopcm));
> > -	ret = intel_vgt_balloon(dev_priv);
> > +	ret = intel_vgt_balloon(i915);
> >   	if (ret)
> >   		return ret;
> > @@ -2891,8 +2891,8 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> >   	/* And finally clear the reserved guard page */
> >   	ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE);
> > -	if (INTEL_PPGTT(dev_priv) == INTEL_PPGTT_ALIASING) {
> > -		ret = init_aliasing_ppgtt(dev_priv);
> > +	if (INTEL_PPGTT(i915) == INTEL_PPGTT_ALIASING) {
> > +		ret = init_aliasing_ppgtt(i915);
> >   		if (ret)
> >   			goto err_appgtt;
> >   	}
> > @@ -2908,18 +2908,18 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv)
> >   /**
> >    * i915_ggtt_cleanup_hw - Clean up GGTT hardware initialization
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    */
> > -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
> > +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	struct i915_vma *vma, *vn;
> >   	struct pagevec *pvec;
> >   	ggtt->vm.closed = true;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	fini_aliasing_ppgtt(dev_priv);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	fini_aliasing_ppgtt(i915);
> >   	list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link)
> >   		WARN_ON(i915_vma_unbind(vma));
> > @@ -2930,24 +2930,24 @@ void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv)
> >   	ggtt_release_guc_top(ggtt);
> >   	if (drm_mm_initialized(&ggtt->vm.mm)) {
> > -		intel_vgt_deballoon(dev_priv);
> > +		intel_vgt_deballoon(i915);
> >   		i915_address_space_fini(&ggtt->vm);
> >   	}
> >   	ggtt->vm.cleanup(&ggtt->vm);
> > -	pvec = &dev_priv->mm.wc_stash.pvec;
> > +	pvec = &i915->mm.wc_stash.pvec;
> >   	if (pvec->nr) {
> >   		set_pages_array_wb(pvec->pages, pvec->nr);
> >   		__pagevec_release(pvec);
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	arch_phys_wc_del(ggtt->mtrr);
> >   	io_mapping_fini(&ggtt->iomap);
> > -	i915_gem_cleanup_stolen(dev_priv);
> > +	i915_gem_cleanup_stolen(i915);
> >   }
> >   static unsigned int gen6_get_total_gtt_size(u16 snb_gmch_ctl)
> > @@ -2986,8 +2986,8 @@ static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl)
> >   static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
> >   {
> > -	struct drm_i915_private *dev_priv = ggtt->vm.i915;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = ggtt->vm.i915;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	phys_addr_t phys_addr;
> >   	int ret;
> > @@ -3001,7 +3001,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
> >   	 * resort to an uncached mapping. The WC issue is easily caught by the
> >   	 * readback check when writing GTT PTE entries.
> >   	 */
> > -	if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> > +	if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10)
> >   		ggtt->gsm = ioremap_nocache(phys_addr, size);
> >   	else
> >   		ggtt->gsm = ioremap_wc(phys_addr, size);
> > @@ -3135,9 +3135,9 @@ void intel_ppat_put(const struct intel_ppat_entry *entry)
> >   	kref_put(&ppat->entries[index].ref, release_ppat);
> >   }
> > -static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv)
> > +static void cnl_private_pat_update_hw(struct drm_i915_private *i915)
> >   {
> > -	struct intel_ppat *ppat = &dev_priv->ppat;
> > +	struct intel_ppat *ppat = &i915->ppat;
> >   	int i;
> >   	for_each_set_bit(i, ppat->dirty, ppat->max_entries) {
> > @@ -3146,9 +3146,9 @@ static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void bdw_private_pat_update_hw(struct drm_i915_private *dev_priv)
> > +static void bdw_private_pat_update_hw(struct drm_i915_private *i915)
> >   {
> > -	struct intel_ppat *ppat = &dev_priv->ppat;
> > +	struct intel_ppat *ppat = &i915->ppat;
> >   	u64 pat = 0;
> >   	int i;
> > @@ -3293,16 +3293,16 @@ static void gen6_gmch_remove(struct i915_address_space *vm)
> >   	cleanup_scratch_page(vm);
> >   }
> > -static void setup_private_pat(struct drm_i915_private *dev_priv)
> > +static void setup_private_pat(struct drm_i915_private *i915)
> >   {
> > -	struct intel_ppat *ppat = &dev_priv->ppat;
> > +	struct intel_ppat *ppat = &i915->ppat;
> >   	int i;
> > -	ppat->i915 = dev_priv;
> > +	ppat->i915 = i915;
> > -	if (INTEL_GEN(dev_priv) >= 10)
> > +	if (INTEL_GEN(i915) >= 10)
> >   		cnl_setup_private_ppat(ppat);
> > -	else if (IS_CHERRYVIEW(dev_priv) || IS_GEN9_LP(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915) || IS_GEN9_LP(i915))
> >   		chv_setup_private_ppat(ppat);
> >   	else
> >   		bdw_setup_private_ppat(ppat);
> > @@ -3315,13 +3315,13 @@ static void setup_private_pat(struct drm_i915_private *dev_priv)
> >   		set_bit(i, ppat->dirty);
> >   	}
> > -	ppat->update_hw(dev_priv);
> > +	ppat->update_hw(i915);
> >   }
> >   static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> >   {
> > -	struct drm_i915_private *dev_priv = ggtt->vm.i915;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = ggtt->vm.i915;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	unsigned int size;
> >   	u16 snb_gmch_ctl;
> >   	int err;
> > @@ -3339,7 +3339,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> >   		DRM_ERROR("Can't set DMA mask/consistent mask (%d)\n", err);
> >   	pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		size = chv_get_total_gtt_size(snb_gmch_ctl);
> >   	else
> >   		size = gen8_get_total_gtt_size(snb_gmch_ctl);
> > @@ -3348,23 +3348,23 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> >   	ggtt->vm.cleanup = gen6_gmch_remove;
> >   	ggtt->vm.insert_page = gen8_ggtt_insert_page;
> >   	ggtt->vm.clear_range = nop_clear_range;
> > -	if (intel_scanout_needs_vtd_wa(dev_priv))
> > +	if (intel_scanout_needs_vtd_wa(i915))
> >   		ggtt->vm.clear_range = gen8_ggtt_clear_range;
> >   	ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
> >   	/* Serialize GTT updates with aperture access on BXT if VT-d is on. */
> > -	if (intel_ggtt_update_needs_vtd_wa(dev_priv) ||
> > -	    IS_CHERRYVIEW(dev_priv) /* fails with concurrent use/update */) {
> > +	if (intel_ggtt_update_needs_vtd_wa(i915) ||
> > +	    IS_CHERRYVIEW(i915) /* fails with concurrent use/update */) {
> >   		ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
> >   		ggtt->vm.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
> >   		if (ggtt->vm.clear_range != nop_clear_range)
> >   			ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL;
> >   		/* Prevent recursively calling stop_machine() and deadlocks. */
> > -		dev_info(dev_priv->drm.dev,
> > +		dev_info(i915->drm.dev,
> >   			 "Disabling error capture for VT-d workaround\n");
> > -		i915_disable_error_state(dev_priv, -ENODEV);
> > +		i915_disable_error_state(i915, -ENODEV);
> >   	}
> >   	ggtt->invalidate = gen6_ggtt_invalidate;
> > @@ -3376,15 +3376,15 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
> >   	ggtt->vm.pte_encode = gen8_pte_encode;
> > -	setup_private_pat(dev_priv);
> > +	setup_private_pat(i915);
> >   	return ggtt_probe_common(ggtt, size);
> >   }
> >   static int gen6_gmch_probe(struct i915_ggtt *ggtt)
> >   {
> > -	struct drm_i915_private *dev_priv = ggtt->vm.i915;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct drm_i915_private *i915 = ggtt->vm.i915;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	unsigned int size;
> >   	u16 snb_gmch_ctl;
> >   	int err;
> > @@ -3413,7 +3413,7 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt)
> >   	ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE;
> >   	ggtt->vm.clear_range = nop_clear_range;
> > -	if (!HAS_FULL_PPGTT(dev_priv) || intel_scanout_needs_vtd_wa(dev_priv))
> > +	if (!HAS_FULL_PPGTT(i915) || intel_scanout_needs_vtd_wa(i915))
> >   		ggtt->vm.clear_range = gen6_ggtt_clear_range;
> >   	ggtt->vm.insert_page = gen6_ggtt_insert_page;
> >   	ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
> > @@ -3421,13 +3421,13 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt)
> >   	ggtt->invalidate = gen6_ggtt_invalidate;
> > -	if (HAS_EDRAM(dev_priv))
> > +	if (HAS_EDRAM(i915))
> >   		ggtt->vm.pte_encode = iris_pte_encode;
> > -	else if (IS_HASWELL(dev_priv))
> > +	else if (IS_HASWELL(i915))
> >   		ggtt->vm.pte_encode = hsw_pte_encode;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		ggtt->vm.pte_encode = byt_pte_encode;
> > -	else if (INTEL_GEN(dev_priv) >= 7)
> > +	else if (INTEL_GEN(i915) >= 7)
> >   		ggtt->vm.pte_encode = ivb_pte_encode;
> >   	else
> >   		ggtt->vm.pte_encode = snb_pte_encode;
> > @@ -3447,11 +3447,11 @@ static void i915_gmch_remove(struct i915_address_space *vm)
> >   static int i915_gmch_probe(struct i915_ggtt *ggtt)
> >   {
> > -	struct drm_i915_private *dev_priv = ggtt->vm.i915;
> > +	struct drm_i915_private *i915 = ggtt->vm.i915;
> >   	phys_addr_t gmadr_base;
> >   	int ret;
> > -	ret = intel_gmch_probe(dev_priv->bridge_dev, dev_priv->drm.pdev, NULL);
> > +	ret = intel_gmch_probe(i915->bridge_dev, i915->drm.pdev, NULL);
> >   	if (!ret) {
> >   		DRM_ERROR("failed to set up gmch\n");
> >   		return -EIO;
> > @@ -3463,7 +3463,7 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt)
> >   		(struct resource) DEFINE_RES_MEM(gmadr_base,
> >   						 ggtt->mappable_end);
> > -	ggtt->do_idle_maps = needs_idle_maps(dev_priv);
> > +	ggtt->do_idle_maps = needs_idle_maps(i915);
> >   	ggtt->vm.insert_page = i915_ggtt_insert_page;
> >   	ggtt->vm.insert_entries = i915_ggtt_insert_entries;
> >   	ggtt->vm.clear_range = i915_ggtt_clear_range;
> > @@ -3484,19 +3484,19 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt)
> >   /**
> >    * i915_ggtt_probe_hw - Probe GGTT hardware location
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    */
> > -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
> > +int i915_ggtt_probe_hw(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	int ret;
> > -	ggtt->vm.i915 = dev_priv;
> > -	ggtt->vm.dma = &dev_priv->drm.pdev->dev;
> > +	ggtt->vm.i915 = i915;
> > +	ggtt->vm.dma = &i915->drm.pdev->dev;
> > -	if (INTEL_GEN(dev_priv) <= 5)
> > +	if (INTEL_GEN(i915) <= 5)
> >   		ret = i915_gmch_probe(ggtt);
> > -	else if (INTEL_GEN(dev_priv) < 8)
> > +	else if (INTEL_GEN(i915) < 8)
> >   		ret = gen6_gmch_probe(ggtt);
> >   	else
> >   		ret = gen8_gmch_probe(ggtt);
> > @@ -3532,35 +3532,35 @@ int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv)
> >   /**
> >    * i915_ggtt_init_hw - Initialize GGTT hardware
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    */
> > -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> > +int i915_ggtt_init_hw(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	int ret;
> > -	stash_init(&dev_priv->mm.wc_stash);
> > +	stash_init(&i915->mm.wc_stash);
> >   	/* Note that we use page colouring to enforce a guard page at the
> >   	 * end of the address space. This is required as the CS may prefetch
> >   	 * beyond the end of the batch buffer, across the page boundary,
> >   	 * and beyond the end of the GTT if we do not provide a guard.
> >   	 */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT);
> >   	ggtt->vm.is_ggtt = true;
> >   	/* Only VLV supports read-only GGTT mappings */
> > -	ggtt->vm.has_read_only = IS_VALLEYVIEW(dev_priv);
> > +	ggtt->vm.has_read_only = IS_VALLEYVIEW(i915);
> > -	if (!HAS_LLC(dev_priv) && !HAS_PPGTT(dev_priv))
> > +	if (!HAS_LLC(i915) && !HAS_PPGTT(i915))
> >   		ggtt->vm.mm.color_adjust = i915_gtt_color_adjust;
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	if (!io_mapping_init_wc(&dev_priv->ggtt.iomap,
> > -				dev_priv->ggtt.gmadr.start,
> > -				dev_priv->ggtt.mappable_end)) {
> > +	if (!io_mapping_init_wc(&i915->ggtt.iomap,
> > +				i915->ggtt.gmadr.start,
> > +				i915->ggtt.mappable_end)) {
> >   		ret = -EIO;
> >   		goto out_gtt_cleanup;
> >   	}
> > @@ -3571,7 +3571,7 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> >   	 * Initialise stolen early so that we may reserve preallocated
> >   	 * objects for the BIOS to KMS transition.
> >   	 */
> > -	ret = i915_gem_init_stolen(dev_priv);
> > +	ret = i915_gem_init_stolen(i915);
> >   	if (ret)
> >   		goto out_gtt_cleanup;
> > @@ -3582,9 +3582,9 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv)
> >   	return ret;
> >   }
> > -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv)
> > +int i915_ggtt_enable_hw(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) < 6 && !intel_enable_gtt())
> > +	if (INTEL_GEN(i915) < 6 && !intel_enable_gtt())
> >   		return -EIO;
> >   	return 0;
> > @@ -3613,12 +3613,12 @@ void i915_ggtt_disable_guc(struct drm_i915_private *i915)
> >   	i915_ggtt_invalidate(i915);
> >   }
> > -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
> > +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	struct i915_vma *vma, *vn;
> > -	i915_check_and_clear_faults(dev_priv);
> > +	i915_check_and_clear_faults(i915);
> >   	mutex_lock(&ggtt->vm.mutex);
> > @@ -3652,15 +3652,15 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv)
> >   	}
> >   	ggtt->vm.closed = false;
> > -	i915_ggtt_invalidate(dev_priv);
> > +	i915_ggtt_invalidate(i915);
> >   	mutex_unlock(&ggtt->vm.mutex);
> > -	if (INTEL_GEN(dev_priv) >= 8) {
> > -		struct intel_ppat *ppat = &dev_priv->ppat;
> > +	if (INTEL_GEN(i915) >= 8) {
> > +		struct intel_ppat *ppat = &i915->ppat;
> >   		bitmap_set(ppat->dirty, 0, ppat->max_entries);
> > -		dev_priv->ppat.update_hw(dev_priv);
> > +		i915->ppat.update_hw(i915);
> >   		return;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
> > index 89437d0a721c..b444552ccc63 100644
> > --- a/drivers/gpu/drm/i915/i915_gem_gtt.h
> > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
> > @@ -400,7 +400,7 @@ struct i915_ggtt {
> >   	/** "Graphics Stolen Memory" holds the global PTEs */
> >   	void __iomem *gsm;
> > -	void (*invalidate)(struct drm_i915_private *dev_priv);
> > +	void (*invalidate)(struct drm_i915_private *i915);
> >   	bool do_idle_maps;
> > @@ -627,17 +627,17 @@ const struct intel_ppat_entry *
> >   intel_ppat_get(struct drm_i915_private *i915, u8 value);
> >   void intel_ppat_put(const struct intel_ppat_entry *entry);
> > -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv);
> > -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv);
> > -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv);
> > +int i915_ggtt_probe_hw(struct drm_i915_private *i915);
> > +int i915_ggtt_init_hw(struct drm_i915_private *i915);
> > +int i915_ggtt_enable_hw(struct drm_i915_private *i915);
> >   void i915_ggtt_enable_guc(struct drm_i915_private *i915);
> >   void i915_ggtt_disable_guc(struct drm_i915_private *i915);
> > -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv);
> > -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv);
> > +int i915_gem_init_ggtt(struct drm_i915_private *i915);
> > +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915);
> > -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv);
> > +int i915_ppgtt_init_hw(struct drm_i915_private *i915);
> > -struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *dev_priv);
> > +struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *i915);
> >   static inline struct i915_address_space *
> >   i915_vm_get(struct i915_address_space *vm)
> > @@ -657,8 +657,8 @@ int gen6_ppgtt_pin(struct i915_ppgtt *base);
> >   void gen6_ppgtt_unpin(struct i915_ppgtt *base);
> >   void gen6_ppgtt_unpin_all(struct i915_ppgtt *base);
> > -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv);
> > -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv);
> > +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915);
> > +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915);
> >   int __must_check i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj,
> >   					    struct sg_table *pages);
> > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
> > index d7f164c9f2eb..335da59031b0 100644
> > --- a/drivers/gpu/drm/i915/i915_gpu_error.c
> > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c
> > @@ -1122,22 +1122,22 @@ static u32 i915_error_generate_code(struct i915_gpu_state *error,
> >   static void gem_record_fences(struct i915_gpu_state *error)
> >   {
> > -	struct drm_i915_private *dev_priv = error->i915;
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = error->i915;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	int i;
> > -	if (INTEL_GEN(dev_priv) >= 6) {
> > -		for (i = 0; i < dev_priv->num_fence_regs; i++)
> > +	if (INTEL_GEN(i915) >= 6) {
> > +		for (i = 0; i < i915->num_fence_regs; i++)
> >   			error->fence[i] =
> >   				intel_uncore_read64(uncore,
> >   						    FENCE_REG_GEN6_LO(i));
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > -		for (i = 0; i < dev_priv->num_fence_regs; i++)
> > +	} else if (INTEL_GEN(i915) >= 4) {
> > +		for (i = 0; i < i915->num_fence_regs; i++)
> >   			error->fence[i] =
> >   				intel_uncore_read64(uncore,
> >   						    FENCE_REG_965_LO(i));
> >   	} else {
> > -		for (i = 0; i < dev_priv->num_fence_regs; i++)
> > +		for (i = 0; i < i915->num_fence_regs; i++)
> >   			error->fence[i] =
> >   				intel_uncore_read(uncore, FENCE_REG(i));
> >   	}
> > @@ -1148,23 +1148,23 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> >   					  struct intel_engine_cs *engine,
> >   					  struct drm_i915_error_engine *ee)
> >   {
> > -	struct drm_i915_private *dev_priv = engine->i915;
> > +	struct drm_i915_private *i915 = engine->i915;
> > -	if (INTEL_GEN(dev_priv) >= 6) {
> > +	if (INTEL_GEN(i915) >= 6) {
> >   		ee->rc_psmi = ENGINE_READ(engine, RING_PSMI_CTL);
> > -		if (INTEL_GEN(dev_priv) >= 8)
> > +		if (INTEL_GEN(i915) >= 8)
> >   			ee->fault_reg = I915_READ(GEN8_RING_FAULT_REG);
> >   		else
> >   			ee->fault_reg = GEN6_RING_FAULT_REG_READ(engine);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		ee->faddr = ENGINE_READ(engine, RING_DMA_FADD);
> >   		ee->ipeir = ENGINE_READ(engine, RING_IPEIR);
> >   		ee->ipehr = ENGINE_READ(engine, RING_IPEHR);
> >   		ee->instps = ENGINE_READ(engine, RING_INSTPS);
> >   		ee->bbaddr = ENGINE_READ(engine, RING_BBADDR);
> > -		if (INTEL_GEN(dev_priv) >= 8) {
> > +		if (INTEL_GEN(i915) >= 8) {
> >   			ee->faddr |= (u64)ENGINE_READ(engine, RING_DMA_FADD_UDW) << 32;
> >   			ee->bbaddr |= (u64)ENGINE_READ(engine, RING_BBADDR_UDW) << 32;
> >   		}
> > @@ -1183,13 +1183,13 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> >   	ee->head = ENGINE_READ(engine, RING_HEAD);
> >   	ee->tail = ENGINE_READ(engine, RING_TAIL);
> >   	ee->ctl = ENGINE_READ(engine, RING_CTL);
> > -	if (INTEL_GEN(dev_priv) > 2)
> > +	if (INTEL_GEN(i915) > 2)
> >   		ee->mode = ENGINE_READ(engine, RING_MI_MODE);
> > -	if (!HWS_NEEDS_PHYSICAL(dev_priv)) {
> > +	if (!HWS_NEEDS_PHYSICAL(i915)) {
> >   		i915_reg_t mmio;
> > -		if (IS_GEN(dev_priv, 7)) {
> > +		if (IS_GEN(i915, 7)) {
> >   			switch (engine->id) {
> >   			default:
> >   				MISSING_CASE(engine->id);
> > @@ -1219,21 +1219,21 @@ static void error_record_engine_registers(struct i915_gpu_state *error,
> >   	ee->idle = intel_engine_is_idle(engine);
> >   	if (!ee->idle)
> >   		ee->hangcheck_timestamp = engine->hangcheck.action_timestamp;
> > -	ee->reset_count = i915_reset_engine_count(&dev_priv->gpu_error,
> > +	ee->reset_count = i915_reset_engine_count(&i915->gpu_error,
> >   						  engine);
> > -	if (HAS_PPGTT(dev_priv)) {
> > +	if (HAS_PPGTT(i915)) {
> >   		int i;
> >   		ee->vm_info.gfx_mode = ENGINE_READ(engine, RING_MODE_GEN7);
> > -		if (IS_GEN(dev_priv, 6)) {
> > +		if (IS_GEN(i915, 6)) {
> >   			ee->vm_info.pp_dir_base =
> >   				ENGINE_READ(engine, RING_PP_DIR_BASE_READ);
> > -		} else if (IS_GEN(dev_priv, 7)) {
> > +		} else if (IS_GEN(i915, 7)) {
> >   			ee->vm_info.pp_dir_base =
> >   				ENGINE_READ(engine, RING_PP_DIR_BASE);
> > -		} else if (INTEL_GEN(dev_priv) >= 8) {
> > +		} else if (INTEL_GEN(i915) >= 8) {
> >   			u32 base = engine->mmio_base;
> >   			for (i = 0; i < 4; i++) {
> > @@ -1388,7 +1388,7 @@ static void request_record_user_bo(struct i915_request *request,
> >   }
> >   static struct drm_i915_error_object *
> > -capture_object(struct drm_i915_private *dev_priv,
> > +capture_object(struct drm_i915_private *i915,
> >   	       struct drm_i915_gem_object *obj)
> >   {
> >   	if (obj && i915_gem_object_has_pages(obj)) {
> > @@ -1399,7 +1399,7 @@ capture_object(struct drm_i915_private *dev_priv,
> >   			.obj = obj,
> >   		};
> > -		return i915_error_object_create(dev_priv, &fake);
> > +		return i915_error_object_create(i915, &fake);
> >   	} else {
> >   		return NULL;
> >   	}
> > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.h b/drivers/gpu/drm/i915/i915_gpu_error.h
> > index 2ecd0c6a1c94..980f95fb5578 100644
> > --- a/drivers/gpu/drm/i915/i915_gpu_error.h
> > +++ b/drivers/gpu/drm/i915/i915_gpu_error.h
> > @@ -234,7 +234,7 @@ struct i915_gpu_error {
> >   	/**
> >   	 * Waitqueue to signal when the reset has completed. Used by clients
> > -	 * that wait for dev_priv->mm.wedged to settle.
> > +	 * that wait for i915->mm.wedged to settle.
> >   	 */
> >   	wait_queue_head_t reset_queue;
> > @@ -259,7 +259,7 @@ __printf(2, 3)
> >   void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...);
> >   struct i915_gpu_state *i915_capture_gpu_state(struct drm_i915_private *i915);
> > -void i915_capture_error_state(struct drm_i915_private *dev_priv,
> > +void i915_capture_error_state(struct drm_i915_private *i915,
> >   			      intel_engine_mask_t engine_mask,
> >   			      const char *error_msg);
> > @@ -286,7 +286,7 @@ void i915_disable_error_state(struct drm_i915_private *i915, int err);
> >   #else
> > -static inline void i915_capture_error_state(struct drm_i915_private *dev_priv,
> > +static inline void i915_capture_error_state(struct drm_i915_private *i915,
> >   					    u32 engine_mask,
> >   					    const char *error_msg)
> >   {
> > diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
> > index cd9edddd6718..81ea65d85b9f 100644
> > --- a/drivers/gpu/drm/i915/i915_irq.c
> > +++ b/drivers/gpu/drm/i915/i915_irq.c
> > @@ -256,18 +256,18 @@ static void gen2_irq_init(struct intel_uncore *uncore,
> >   #define GEN2_IRQ_INIT(uncore, imr_val, ier_val) \
> >   	gen2_irq_init((uncore), imr_val, ier_val)
> > -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
> > -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir);
> > +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir);
> > +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 pm_iir);
> >   /* For display hotplug interrupt */
> >   static inline void
> > -i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
> > +i915_hotplug_interrupt_update_locked(struct drm_i915_private *i915,
> >   				     u32 mask,
> >   				     u32 bits)
> >   {
> >   	u32 val;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	WARN_ON(bits & ~mask);
> >   	val = I915_READ(PORT_HOTPLUG_EN);
> > @@ -278,7 +278,7 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
> >   /**
> >    * i915_hotplug_interrupt_update - update hotplug interrupt enable
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @mask: bits to update
> >    * @bits: bits to enable
> >    * NOTE: the HPD enable bits are modified both inside and outside
> > @@ -288,13 +288,13 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv,
> >    * held already, this function acquires the lock itself. A non-locking
> >    * version is also available.
> >    */
> > -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv,
> > +void i915_hotplug_interrupt_update(struct drm_i915_private *i915,
> >   				   u32 mask,
> >   				   u32 bits)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	i915_hotplug_interrupt_update_locked(dev_priv, mask, bits);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915_hotplug_interrupt_update_locked(i915, mask, bits);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> >   static u32
> > @@ -334,84 +334,84 @@ static bool gen11_reset_one_iir(struct drm_i915_private * const i915,
> >   /**
> >    * ilk_update_display_irq - update DEIMR
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -void ilk_update_display_irq(struct drm_i915_private *dev_priv,
> > +void ilk_update_display_irq(struct drm_i915_private *i915,
> >   			    u32 interrupt_mask,
> >   			    u32 enabled_irq_mask)
> >   {
> >   	u32 new_val;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> > -	new_val = dev_priv->irq_mask;
> > +	new_val = i915->irq_mask;
> >   	new_val &= ~interrupt_mask;
> >   	new_val |= (~enabled_irq_mask & interrupt_mask);
> > -	if (new_val != dev_priv->irq_mask) {
> > -		dev_priv->irq_mask = new_val;
> > -		I915_WRITE(DEIMR, dev_priv->irq_mask);
> > +	if (new_val != i915->irq_mask) {
> > +		i915->irq_mask = new_val;
> > +		I915_WRITE(DEIMR, i915->irq_mask);
> >   		POSTING_READ(DEIMR);
> >   	}
> >   }
> >   /**
> >    * ilk_update_gt_irq - update GTIMR
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -static void ilk_update_gt_irq(struct drm_i915_private *dev_priv,
> > +static void ilk_update_gt_irq(struct drm_i915_private *i915,
> >   			      u32 interrupt_mask,
> >   			      u32 enabled_irq_mask)
> >   {
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> > -	dev_priv->gt_irq_mask &= ~interrupt_mask;
> > -	dev_priv->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask);
> > -	I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
> > +	i915->gt_irq_mask &= ~interrupt_mask;
> > +	i915->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask);
> > +	I915_WRITE(GTIMR, i915->gt_irq_mask);
> >   }
> > -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask)
> > +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask)
> >   {
> > -	ilk_update_gt_irq(dev_priv, mask, mask);
> > -	intel_uncore_posting_read_fw(&dev_priv->uncore, GTIMR);
> > +	ilk_update_gt_irq(i915, mask, mask);
> > +	intel_uncore_posting_read_fw(&i915->uncore, GTIMR);
> >   }
> > -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask)
> > +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask)
> >   {
> > -	ilk_update_gt_irq(dev_priv, mask, 0);
> > +	ilk_update_gt_irq(i915, mask, 0);
> >   }
> > -static i915_reg_t gen6_pm_iir(struct drm_i915_private *dev_priv)
> > +static i915_reg_t gen6_pm_iir(struct drm_i915_private *i915)
> >   {
> > -	WARN_ON_ONCE(INTEL_GEN(dev_priv) >= 11);
> > +	WARN_ON_ONCE(INTEL_GEN(i915) >= 11);
> > -	return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR;
> > +	return INTEL_GEN(i915) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR;
> >   }
> > -static void write_pm_imr(struct drm_i915_private *dev_priv)
> > +static void write_pm_imr(struct drm_i915_private *i915)
> >   {
> >   	i915_reg_t reg;
> > -	u32 mask = dev_priv->pm_imr;
> > +	u32 mask = i915->pm_imr;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		reg = GEN11_GPM_WGBOXPERF_INTR_MASK;
> >   		/* pm is in upper half */
> >   		mask = mask << 16;
> > -	} else if (INTEL_GEN(dev_priv) >= 8) {
> > +	} else if (INTEL_GEN(i915) >= 8) {
> >   		reg = GEN8_GT_IMR(2);
> >   	} else {
> >   		reg = GEN6_PMIMR;
> > @@ -421,16 +421,16 @@ static void write_pm_imr(struct drm_i915_private *dev_priv)
> >   	POSTING_READ(reg);
> >   }
> > -static void write_pm_ier(struct drm_i915_private *dev_priv)
> > +static void write_pm_ier(struct drm_i915_private *i915)
> >   {
> >   	i915_reg_t reg;
> > -	u32 mask = dev_priv->pm_ier;
> > +	u32 mask = i915->pm_ier;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		reg = GEN11_GPM_WGBOXPERF_INTR_ENABLE;
> >   		/* pm is in upper half */
> >   		mask = mask << 16;
> > -	} else if (INTEL_GEN(dev_priv) >= 8) {
> > +	} else if (INTEL_GEN(i915) >= 8) {
> >   		reg = GEN8_GT_IER(2);
> >   	} else {
> >   		reg = GEN6_PMIER;
> > @@ -441,11 +441,11 @@ static void write_pm_ier(struct drm_i915_private *dev_priv)
> >   /**
> >    * snb_update_pm_irq - update GEN6_PMIMR
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -static void snb_update_pm_irq(struct drm_i915_private *dev_priv,
> > +static void snb_update_pm_irq(struct drm_i915_private *i915,
> >   			      u32 interrupt_mask,
> >   			      u32 enabled_irq_mask)
> >   {
> > @@ -453,127 +453,127 @@ static void snb_update_pm_irq(struct drm_i915_private *dev_priv,
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	new_val = dev_priv->pm_imr;
> > +	new_val = i915->pm_imr;
> >   	new_val &= ~interrupt_mask;
> >   	new_val |= (~enabled_irq_mask & interrupt_mask);
> > -	if (new_val != dev_priv->pm_imr) {
> > -		dev_priv->pm_imr = new_val;
> > -		write_pm_imr(dev_priv);
> > +	if (new_val != i915->pm_imr) {
> > +		i915->pm_imr = new_val;
> > +		write_pm_imr(i915);
> >   	}
> >   }
> > -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> > +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask)
> >   {
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> > -	snb_update_pm_irq(dev_priv, mask, mask);
> > +	snb_update_pm_irq(i915, mask, mask);
> >   }
> > -static void __gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> > +static void __gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask)
> >   {
> > -	snb_update_pm_irq(dev_priv, mask, 0);
> > +	snb_update_pm_irq(i915, mask, 0);
> >   }
> > -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask)
> > +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask)
> >   {
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> > -	__gen6_mask_pm_irq(dev_priv, mask);
> > +	__gen6_mask_pm_irq(i915, mask);
> >   }
> > -static void gen6_reset_pm_iir(struct drm_i915_private *dev_priv, u32 reset_mask)
> > +static void gen6_reset_pm_iir(struct drm_i915_private *i915, u32 reset_mask)
> >   {
> > -	i915_reg_t reg = gen6_pm_iir(dev_priv);
> > +	i915_reg_t reg = gen6_pm_iir(i915);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	I915_WRITE(reg, reset_mask);
> >   	I915_WRITE(reg, reset_mask);
> >   	POSTING_READ(reg);
> >   }
> > -static void gen6_enable_pm_irq(struct drm_i915_private *dev_priv, u32 enable_mask)
> > +static void gen6_enable_pm_irq(struct drm_i915_private *i915, u32 enable_mask)
> >   {
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	dev_priv->pm_ier |= enable_mask;
> > -	write_pm_ier(dev_priv);
> > -	gen6_unmask_pm_irq(dev_priv, enable_mask);
> > +	i915->pm_ier |= enable_mask;
> > +	write_pm_ier(i915);
> > +	gen6_unmask_pm_irq(i915, enable_mask);
> >   	/* unmask_pm_irq provides an implicit barrier (POSTING_READ) */
> >   }
> > -static void gen6_disable_pm_irq(struct drm_i915_private *dev_priv, u32 disable_mask)
> > +static void gen6_disable_pm_irq(struct drm_i915_private *i915, u32 disable_mask)
> >   {
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	dev_priv->pm_ier &= ~disable_mask;
> > -	__gen6_mask_pm_irq(dev_priv, disable_mask);
> > -	write_pm_ier(dev_priv);
> > +	i915->pm_ier &= ~disable_mask;
> > +	__gen6_mask_pm_irq(i915, disable_mask);
> > +	write_pm_ier(i915);
> >   	/* though a barrier is missing here, but don't really need a one */
> >   }
> > -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv)
> > +void gen11_reset_rps_interrupts(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	while (gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM))
> > +	while (gen11_reset_one_iir(i915, 0, GEN11_GTPM))
> >   		;
> > -	dev_priv->gt_pm.rps.pm_iir = 0;
> > +	i915->gt_pm.rps.pm_iir = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv)
> > +void gen6_reset_rps_interrupts(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	gen6_reset_pm_iir(dev_priv, GEN6_PM_RPS_EVENTS);
> > -	dev_priv->gt_pm.rps.pm_iir = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	gen6_reset_pm_iir(i915, GEN6_PM_RPS_EVENTS);
> > +	i915->gt_pm.rps.pm_iir = 0;
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv)
> > +void gen6_enable_rps_interrupts(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	if (READ_ONCE(rps->interrupts_enabled))
> >   		return;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	WARN_ON_ONCE(rps->pm_iir);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM));
> > +	if (INTEL_GEN(i915) >= 11)
> > +		WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GTPM));
> >   	else
> > -		WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) & dev_priv->pm_rps_events);
> > +		WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) & i915->pm_rps_events);
> >   	rps->interrupts_enabled = true;
> > -	gen6_enable_pm_irq(dev_priv, dev_priv->pm_rps_events);
> > +	gen6_enable_pm_irq(i915, i915->pm_rps_events);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
> > +void gen6_disable_rps_interrupts(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	if (!READ_ONCE(rps->interrupts_enabled))
> >   		return;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	rps->interrupts_enabled = false;
> > -	I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u));
> > +	I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(i915, ~0u));
> > -	gen6_disable_pm_irq(dev_priv, GEN6_PM_RPS_EVENTS);
> > +	gen6_disable_pm_irq(i915, GEN6_PM_RPS_EVENTS);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	spin_unlock_irq(&i915->irq_lock);
> > +	synchronize_irq(i915->drm.irq);
> >   	/* Now that we will not be generating any more work, flush any
> >   	 * outstanding tasks. As we are called on the RPS idle path,
> > @@ -581,48 +581,48 @@ void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv)
> >   	 * state of the worker can be discarded.
> >   	 */
> >   	cancel_work_sync(&rps->work);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		gen11_reset_rps_interrupts(dev_priv);
> > +	if (INTEL_GEN(i915) >= 11)
> > +		gen11_reset_rps_interrupts(i915);
> >   	else
> > -		gen6_reset_rps_interrupts(dev_priv);
> > +		gen6_reset_rps_interrupts(i915);
> >   }
> > -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv)
> > +void gen9_reset_guc_interrupts(struct drm_i915_private *i915)
> >   {
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	assert_rpm_wakelock_held(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	gen6_reset_pm_iir(dev_priv, dev_priv->pm_guc_events);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	gen6_reset_pm_iir(i915, i915->pm_guc_events);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv)
> > +void gen9_enable_guc_interrupts(struct drm_i915_private *i915)
> >   {
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	assert_rpm_wakelock_held(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (!dev_priv->guc.interrupts.enabled) {
> > -		WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) &
> > -				       dev_priv->pm_guc_events);
> > -		dev_priv->guc.interrupts.enabled = true;
> > -		gen6_enable_pm_irq(dev_priv, dev_priv->pm_guc_events);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (!i915->guc.interrupts.enabled) {
> > +		WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) &
> > +				       i915->pm_guc_events);
> > +		i915->guc.interrupts.enabled = true;
> > +		gen6_enable_pm_irq(i915, i915->pm_guc_events);
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv)
> > +void gen9_disable_guc_interrupts(struct drm_i915_private *i915)
> >   {
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	assert_rpm_wakelock_held(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	dev_priv->guc.interrupts.enabled = false;
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915->guc.interrupts.enabled = false;
> > -	gen6_disable_pm_irq(dev_priv, dev_priv->pm_guc_events);
> > +	gen6_disable_pm_irq(i915, i915->pm_guc_events);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	spin_unlock_irq(&i915->irq_lock);
> > +	synchronize_irq(i915->drm.irq);
> > -	gen9_reset_guc_interrupts(dev_priv);
> > +	gen9_reset_guc_interrupts(i915);
> >   }
> >   void gen11_reset_guc_interrupts(struct drm_i915_private *i915)
> > @@ -632,53 +632,53 @@ void gen11_reset_guc_interrupts(struct drm_i915_private *i915)
> >   	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen11_enable_guc_interrupts(struct drm_i915_private *dev_priv)
> > +void gen11_enable_guc_interrupts(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (!dev_priv->guc.interrupts.enabled) {
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (!i915->guc.interrupts.enabled) {
> >   		u32 events = REG_FIELD_PREP(ENGINE1_MASK,
> >   					    GEN11_GUC_INTR_GUC2HOST);
> > -		WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GUC));
> > +		WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GUC));
> >   		I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, events);
> >   		I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~events);
> > -		dev_priv->guc.interrupts.enabled = true;
> > +		i915->guc.interrupts.enabled = true;
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen11_disable_guc_interrupts(struct drm_i915_private *dev_priv)
> > +void gen11_disable_guc_interrupts(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	dev_priv->guc.interrupts.enabled = false;
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915->guc.interrupts.enabled = false;
> >   	I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~0);
> >   	I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, 0);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	spin_unlock_irq(&i915->irq_lock);
> > +	synchronize_irq(i915->drm.irq);
> > -	gen11_reset_guc_interrupts(dev_priv);
> > +	gen11_reset_guc_interrupts(i915);
> >   }
> >   /**
> >    * bdw_update_port_irq - update DE port interrupt
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -static void bdw_update_port_irq(struct drm_i915_private *dev_priv,
> > +static void bdw_update_port_irq(struct drm_i915_private *i915,
> >   				u32 interrupt_mask,
> >   				u32 enabled_irq_mask)
> >   {
> >   	u32 new_val;
> >   	u32 old_val;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> >   	old_val = I915_READ(GEN8_DE_PORT_IMR);
> > @@ -695,43 +695,43 @@ static void bdw_update_port_irq(struct drm_i915_private *dev_priv,
> >   /**
> >    * bdw_update_pipe_irq - update DE pipe interrupt
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @pipe: pipe whose interrupt to update
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv,
> > +void bdw_update_pipe_irq(struct drm_i915_private *i915,
> >   			 enum pipe pipe,
> >   			 u32 interrupt_mask,
> >   			 u32 enabled_irq_mask)
> >   {
> >   	u32 new_val;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> > -	new_val = dev_priv->de_irq_mask[pipe];
> > +	new_val = i915->de_irq_mask[pipe];
> >   	new_val &= ~interrupt_mask;
> >   	new_val |= (~enabled_irq_mask & interrupt_mask);
> > -	if (new_val != dev_priv->de_irq_mask[pipe]) {
> > -		dev_priv->de_irq_mask[pipe] = new_val;
> > -		I915_WRITE(GEN8_DE_PIPE_IMR(pipe), dev_priv->de_irq_mask[pipe]);
> > +	if (new_val != i915->de_irq_mask[pipe]) {
> > +		i915->de_irq_mask[pipe] = new_val;
> > +		I915_WRITE(GEN8_DE_PIPE_IMR(pipe), i915->de_irq_mask[pipe]);
> >   		POSTING_READ(GEN8_DE_PIPE_IMR(pipe));
> >   	}
> >   }
> >   /**
> >    * ibx_display_interrupt_update - update SDEIMR
> > - * @dev_priv: driver private
> > + * @i915: driver private
> >    * @interrupt_mask: mask of interrupt bits to update
> >    * @enabled_irq_mask: mask of interrupt bits to enable
> >    */
> > -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
> > +void ibx_display_interrupt_update(struct drm_i915_private *i915,
> >   				  u32 interrupt_mask,
> >   				  u32 enabled_irq_mask)
> >   {
> > @@ -741,24 +741,24 @@ void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
> >   	WARN_ON(enabled_irq_mask & ~interrupt_mask);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	if (WARN_ON(!intel_irqs_enabled(dev_priv)))
> > +	if (WARN_ON(!intel_irqs_enabled(i915)))
> >   		return;
> >   	I915_WRITE(SDEIMR, sdeimr);
> >   	POSTING_READ(SDEIMR);
> >   }
> > -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> > +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915,
> >   			      enum pipe pipe)
> >   {
> > -	u32 status_mask = dev_priv->pipestat_irq_mask[pipe];
> > +	u32 status_mask = i915->pipestat_irq_mask[pipe];
> >   	u32 enable_mask = status_mask << 16;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	if (INTEL_GEN(dev_priv) < 5)
> > +	if (INTEL_GEN(i915) < 5)
> >   		goto out;
> >   	/*
> > @@ -791,7 +791,7 @@ u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> >   	return enable_mask;
> >   }
> > -void i915_enable_pipestat(struct drm_i915_private *dev_priv,
> > +void i915_enable_pipestat(struct drm_i915_private *i915,
> >   			  enum pipe pipe, u32 status_mask)
> >   {
> >   	i915_reg_t reg = PIPESTAT(pipe);
> > @@ -801,20 +801,20 @@ void i915_enable_pipestat(struct drm_i915_private *dev_priv,
> >   		  "pipe %c: status_mask=0x%x\n",
> >   		  pipe_name(pipe), status_mask);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > -	WARN_ON(!intel_irqs_enabled(dev_priv));
> > +	lockdep_assert_held(&i915->irq_lock);
> > +	WARN_ON(!intel_irqs_enabled(i915));
> > -	if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == status_mask)
> > +	if ((i915->pipestat_irq_mask[pipe] & status_mask) == status_mask)
> >   		return;
> > -	dev_priv->pipestat_irq_mask[pipe] |= status_mask;
> > -	enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> > +	i915->pipestat_irq_mask[pipe] |= status_mask;
> > +	enable_mask = i915_pipestat_enable_mask(i915, pipe);
> >   	I915_WRITE(reg, enable_mask | status_mask);
> >   	POSTING_READ(reg);
> >   }
> > -void i915_disable_pipestat(struct drm_i915_private *dev_priv,
> > +void i915_disable_pipestat(struct drm_i915_private *i915,
> >   			   enum pipe pipe, u32 status_mask)
> >   {
> >   	i915_reg_t reg = PIPESTAT(pipe);
> > @@ -824,44 +824,44 @@ void i915_disable_pipestat(struct drm_i915_private *dev_priv,
> >   		  "pipe %c: status_mask=0x%x\n",
> >   		  pipe_name(pipe), status_mask);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > -	WARN_ON(!intel_irqs_enabled(dev_priv));
> > +	lockdep_assert_held(&i915->irq_lock);
> > +	WARN_ON(!intel_irqs_enabled(i915));
> > -	if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == 0)
> > +	if ((i915->pipestat_irq_mask[pipe] & status_mask) == 0)
> >   		return;
> > -	dev_priv->pipestat_irq_mask[pipe] &= ~status_mask;
> > -	enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> > +	i915->pipestat_irq_mask[pipe] &= ~status_mask;
> > +	enable_mask = i915_pipestat_enable_mask(i915, pipe);
> >   	I915_WRITE(reg, enable_mask | status_mask);
> >   	POSTING_READ(reg);
> >   }
> > -static bool i915_has_asle(struct drm_i915_private *dev_priv)
> > +static bool i915_has_asle(struct drm_i915_private *i915)
> >   {
> > -	if (!dev_priv->opregion.asle)
> > +	if (!i915->opregion.asle)
> >   		return false;
> > -	return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> > +	return IS_PINEVIEW(i915) || IS_MOBILE(i915);
> >   }
> >   /**
> >    * i915_enable_asle_pipestat - enable ASLE pipestat for OpRegion
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    */
> > -static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv)
> > +static void i915_enable_asle_pipestat(struct drm_i915_private *i915)
> >   {
> > -	if (!i915_has_asle(dev_priv))
> > +	if (!i915_has_asle(i915))
> >   		return;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	i915_enable_pipestat(dev_priv, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > -		i915_enable_pipestat(dev_priv, PIPE_A,
> > +	i915_enable_pipestat(i915, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS);
> > +	if (INTEL_GEN(i915) >= 4)
> > +		i915_enable_pipestat(i915, PIPE_A,
> >   				     PIPE_LEGACY_BLC_EVENT_STATUS);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> >   /*
> > @@ -919,7 +919,7 @@ static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv)
> >    */
> >   static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_vblank_crtc *vblank = &dev->vblank[pipe];
> >   	const struct drm_display_mode *mode = &vblank->hwmode;
> >   	i915_reg_t high_frame, low_frame;
> > @@ -955,7 +955,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >   	high_frame = PIPEFRAME(pipe);
> >   	low_frame = PIPEFRAMEPIXEL(pipe);
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	/*
> >   	 * High & low register fields aren't synchronized, so make sure
> > @@ -968,7 +968,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >   		high2 = I915_READ_FW(high_frame) & PIPE_FRAME_HIGH_MASK;
> >   	} while (high1 != high2);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   	high1 >>= PIPE_FRAME_HIGH_SHIFT;
> >   	pixel = low & PIPE_PIXEL_MASK;
> > @@ -984,7 +984,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >   static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	return I915_READ(PIPE_FRMCOUNT_G4X(pipe));
> >   }
> > @@ -999,7 +999,7 @@ static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe)
> >    */
> >   static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct drm_vblank_crtc *vblank =
> >   		&crtc->base.dev->vblank[drm_crtc_index(&crtc->base)];
> >   	const struct drm_display_mode *mode = &vblank->hwmode;
> > @@ -1044,7 +1044,7 @@ static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc)
> >   static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	const struct drm_display_mode *mode;
> >   	struct drm_vblank_crtc *vblank;
> >   	enum pipe pipe = crtc->pipe;
> > @@ -1063,7 +1063,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> >   	if (mode->flags & DRM_MODE_FLAG_INTERLACE)
> >   		vtotal /= 2;
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN2;
> >   	else
> >   		position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN3;
> > @@ -1080,7 +1080,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc)
> >   	 * problem.  We may need to extend this to include other platforms,
> >   	 * but so far testing only shows the problem on HSW.
> >   	 */
> > -	if (HAS_DDI(dev_priv) && !position) {
> > +	if (HAS_DDI(i915) && !position) {
> >   		int i, temp;
> >   		for (i = 0; i < 100; i++) {
> > @@ -1105,14 +1105,14 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> >   				     ktime_t *stime, ktime_t *etime,
> >   				     const struct drm_display_mode *mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(dev_priv,
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(i915,
> >   								pipe);
> >   	int position;
> >   	int vbl_start, vbl_end, hsync_start, htotal, vtotal;
> >   	unsigned long irqflags;
> > -	bool use_scanline_counter = INTEL_GEN(dev_priv) >= 5 ||
> > -		IS_G4X(dev_priv) || IS_GEN(dev_priv, 2) ||
> > +	bool use_scanline_counter = INTEL_GEN(i915) >= 5 ||
> > +		IS_G4X(i915) || IS_GEN(i915, 2) ||
> >   		mode->private_flags & I915_MODE_FLAG_USE_SCANLINE_COUNTER;
> >   	if (WARN_ON(!mode->crtc_clock)) {
> > @@ -1138,7 +1138,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> >   	 * register reads, potentially with preemption disabled, so the
> >   	 * following code must not block on uncore.lock.
> >   	 */
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	/* preempt_disable_rt() should go right here in PREEMPT_RT patchset. */
> > @@ -1193,7 +1193,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> >   	/* preempt_enable_rt() should go right here in PREEMPT_RT patchset. */
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   	/*
> >   	 * While in vblank, position will be negative
> > @@ -1219,20 +1219,20 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe,
> >   int intel_get_crtc_scanline(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	unsigned long irqflags;
> >   	int position;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	position = __intel_get_crtc_scanline(crtc);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   	return position;
> >   }
> > -static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
> > +static void ironlake_rps_change_irq_handler(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 busy_up, busy_down, max_avg, min_avg;
> >   	u8 new_delay;
> > @@ -1242,7 +1242,7 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
> >   			     MEMINTRSTS,
> >   			     intel_uncore_read(uncore, MEMINTRSTS));
> > -	new_delay = dev_priv->ips.cur_delay;
> > +	new_delay = i915->ips.cur_delay;
> >   	intel_uncore_write16(uncore, MEMINTRSTS, MEMINT_EVAL_CHG);
> >   	busy_up = intel_uncore_read(uncore, RCPREVBSYTUPAVG);
> > @@ -1252,26 +1252,26 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv)
> >   	/* Handle RCS change request from hw */
> >   	if (busy_up > max_avg) {
> > -		if (dev_priv->ips.cur_delay != dev_priv->ips.max_delay)
> > -			new_delay = dev_priv->ips.cur_delay - 1;
> > -		if (new_delay < dev_priv->ips.max_delay)
> > -			new_delay = dev_priv->ips.max_delay;
> > +		if (i915->ips.cur_delay != i915->ips.max_delay)
> > +			new_delay = i915->ips.cur_delay - 1;
> > +		if (new_delay < i915->ips.max_delay)
> > +			new_delay = i915->ips.max_delay;
> >   	} else if (busy_down < min_avg) {
> > -		if (dev_priv->ips.cur_delay != dev_priv->ips.min_delay)
> > -			new_delay = dev_priv->ips.cur_delay + 1;
> > -		if (new_delay > dev_priv->ips.min_delay)
> > -			new_delay = dev_priv->ips.min_delay;
> > +		if (i915->ips.cur_delay != i915->ips.min_delay)
> > +			new_delay = i915->ips.cur_delay + 1;
> > +		if (new_delay > i915->ips.min_delay)
> > +			new_delay = i915->ips.min_delay;
> >   	}
> > -	if (ironlake_set_drps(dev_priv, new_delay))
> > -		dev_priv->ips.cur_delay = new_delay;
> > +	if (ironlake_set_drps(i915, new_delay))
> > +		i915->ips.cur_delay = new_delay;
> >   	spin_unlock(&mchdev_lock);
> >   	return;
> >   }
> > -static void vlv_c0_read(struct drm_i915_private *dev_priv,
> > +static void vlv_c0_read(struct drm_i915_private *i915,
> >   			struct intel_rps_ei *ei)
> >   {
> >   	ei->ktime = ktime_get_raw();
> > @@ -1279,14 +1279,14 @@ static void vlv_c0_read(struct drm_i915_private *dev_priv,
> >   	ei->media_c0 = I915_READ(VLV_MEDIA_C0_COUNT);
> >   }
> > -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv)
> > +void gen6_rps_reset_ei(struct drm_i915_private *i915)
> >   {
> > -	memset(&dev_priv->gt_pm.rps.ei, 0, sizeof(dev_priv->gt_pm.rps.ei));
> > +	memset(&i915->gt_pm.rps.ei, 0, sizeof(i915->gt_pm.rps.ei));
> >   }
> > -static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> > +static u32 vlv_wa_c0_ei(struct drm_i915_private *i915, u32 pm_iir)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	const struct intel_rps_ei *prev = &rps->ei;
> >   	struct intel_rps_ei now;
> >   	u32 events = 0;
> > @@ -1294,7 +1294,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> >   	if ((pm_iir & GEN6_PM_RP_UP_EI_EXPIRED) == 0)
> >   		return 0;
> > -	vlv_c0_read(dev_priv, &now);
> > +	vlv_c0_read(i915, &now);
> >   	if (prev->ktime) {
> >   		u64 time, c0;
> > @@ -1302,7 +1302,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> >   		time = ktime_us_delta(now.ktime, prev->ktime);
> > -		time *= dev_priv->czclk_freq;
> > +		time *= i915->czclk_freq;
> >   		/* Workload can be split between render + media,
> >   		 * e.g. SwapBuffers being blitted in X after being rendered in
> > @@ -1326,28 +1326,28 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir)
> >   static void gen6_pm_rps_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private, gt_pm.rps.work);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	bool client_boost = false;
> >   	int new_delay, adj, min, max;
> >   	u32 pm_iir = 0;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	if (rps->interrupts_enabled) {
> >   		pm_iir = fetch_and_zero(&rps->pm_iir);
> >   		client_boost = atomic_read(&rps->num_waiters);
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/* Make sure we didn't queue anything we're not going to process. */
> > -	WARN_ON(pm_iir & ~dev_priv->pm_rps_events);
> > -	if ((pm_iir & dev_priv->pm_rps_events) == 0 && !client_boost)
> > +	WARN_ON(pm_iir & ~i915->pm_rps_events);
> > +	if ((pm_iir & i915->pm_rps_events) == 0 && !client_boost)
> >   		goto out;
> >   	mutex_lock(&rps->lock);
> > -	pm_iir |= vlv_wa_c0_ei(dev_priv, pm_iir);
> > +	pm_iir |= vlv_wa_c0_ei(i915, pm_iir);
> >   	adj = rps->last_adj;
> >   	new_delay = rps->cur_freq;
> > @@ -1362,7 +1362,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> >   		if (adj > 0)
> >   			adj *= 2;
> >   		else /* CHV needs even encode values */
> > -			adj = IS_CHERRYVIEW(dev_priv) ? 2 : 1;
> > +			adj = IS_CHERRYVIEW(i915) ? 2 : 1;
> >   		if (new_delay >= rps->max_freq_softlimit)
> >   			adj = 0;
> > @@ -1378,7 +1378,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> >   		if (adj < 0)
> >   			adj *= 2;
> >   		else /* CHV needs even encode values */
> > -			adj = IS_CHERRYVIEW(dev_priv) ? -2 : -1;
> > +			adj = IS_CHERRYVIEW(i915) ? -2 : -1;
> >   		if (new_delay <= rps->min_freq_softlimit)
> >   			adj = 0;
> > @@ -1406,7 +1406,7 @@ static void gen6_pm_rps_work(struct work_struct *work)
> >   	new_delay += adj;
> >   	new_delay = clamp_t(int, new_delay, min, max);
> > -	if (intel_set_rps(dev_priv, new_delay)) {
> > +	if (intel_set_rps(i915, new_delay)) {
> >   		DRM_DEBUG_DRIVER("Failed to set new GPU frequency\n");
> >   		rps->last_adj = 0;
> >   	}
> > @@ -1415,10 +1415,10 @@ static void gen6_pm_rps_work(struct work_struct *work)
> >   out:
> >   	/* Make sure not to corrupt PMIMR state used by ringbuffer on GEN6 */
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	if (rps->interrupts_enabled)
> > -		gen6_unmask_pm_irq(dev_priv, dev_priv->pm_rps_events);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +		gen6_unmask_pm_irq(i915, i915->pm_rps_events);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > @@ -1433,8 +1433,8 @@ static void gen6_pm_rps_work(struct work_struct *work)
> >    */
> >   static void ivybridge_parity_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv), l3_parity.error_work);
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915), l3_parity.error_work);
> >   	u32 error_status, row, bank, subbank;
> >   	char *parity_event[6];
> >   	u32 misccpctl;
> > @@ -1444,24 +1444,24 @@ static void ivybridge_parity_work(struct work_struct *work)
> >   	 * In order to prevent a get/put style interface, acquire struct mutex
> >   	 * any time we access those registers.
> >   	 */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	/* If we've screwed up tracking, just let the interrupt fire again */
> > -	if (WARN_ON(!dev_priv->l3_parity.which_slice))
> > +	if (WARN_ON(!i915->l3_parity.which_slice))
> >   		goto out;
> >   	misccpctl = I915_READ(GEN7_MISCCPCTL);
> >   	I915_WRITE(GEN7_MISCCPCTL, misccpctl & ~GEN7_DOP_CLOCK_GATE_ENABLE);
> >   	POSTING_READ(GEN7_MISCCPCTL);
> > -	while ((slice = ffs(dev_priv->l3_parity.which_slice)) != 0) {
> > +	while ((slice = ffs(i915->l3_parity.which_slice)) != 0) {
> >   		i915_reg_t reg;
> >   		slice--;
> > -		if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(dev_priv)))
> > +		if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(i915)))
> >   			break;
> > -		dev_priv->l3_parity.which_slice &= ~(1<<slice);
> > +		i915->l3_parity.which_slice &= ~(1<<slice);
> >   		reg = GEN7_L3CDERRST1(slice);
> > @@ -1480,7 +1480,7 @@ static void ivybridge_parity_work(struct work_struct *work)
> >   		parity_event[4] = kasprintf(GFP_KERNEL, "SLICE=%d", slice);
> >   		parity_event[5] = NULL;
> > -		kobject_uevent_env(&dev_priv->drm.primary->kdev->kobj,
> > +		kobject_uevent_env(&i915->drm.primary->kdev->kobj,
> >   				   KOBJ_CHANGE, parity_event);
> >   		DRM_DEBUG("Parity error: Slice = %d, Row = %d, Bank = %d, Sub bank = %d.\n",
> > @@ -1495,60 +1495,60 @@ static void ivybridge_parity_work(struct work_struct *work)
> >   	I915_WRITE(GEN7_MISCCPCTL, misccpctl);
> >   out:
> > -	WARN_ON(dev_priv->l3_parity.which_slice);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	gen5_enable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	WARN_ON(i915->l3_parity.which_slice);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	gen5_enable_gt_irq(i915, GT_PARITY_ERROR(i915));
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   }
> > -static void ivybridge_parity_error_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ivybridge_parity_error_irq_handler(struct drm_i915_private *i915,
> >   					       u32 iir)
> >   {
> > -	if (!HAS_L3_DPF(dev_priv))
> > +	if (!HAS_L3_DPF(i915))
> >   		return;
> > -	spin_lock(&dev_priv->irq_lock);
> > -	gen5_disable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv));
> > -	spin_unlock(&dev_priv->irq_lock);
> > +	spin_lock(&i915->irq_lock);
> > +	gen5_disable_gt_irq(i915, GT_PARITY_ERROR(i915));
> > +	spin_unlock(&i915->irq_lock);
> > -	iir &= GT_PARITY_ERROR(dev_priv);
> > +	iir &= GT_PARITY_ERROR(i915);
> >   	if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1)
> > -		dev_priv->l3_parity.which_slice |= 1 << 1;
> > +		i915->l3_parity.which_slice |= 1 << 1;
> >   	if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT)
> > -		dev_priv->l3_parity.which_slice |= 1 << 0;
> > +		i915->l3_parity.which_slice |= 1 << 0;
> > -	queue_work(dev_priv->wq, &dev_priv->l3_parity.error_work);
> > +	queue_work(i915->wq, &i915->l3_parity.error_work);
> >   }
> > -static void ilk_gt_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ilk_gt_irq_handler(struct drm_i915_private *i915,
> >   			       u32 gt_iir)
> >   {
> >   	if (gt_iir & GT_RENDER_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> >   	if (gt_iir & ILK_BSD_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
> >   }
> > -static void snb_gt_irq_handler(struct drm_i915_private *dev_priv,
> > +static void snb_gt_irq_handler(struct drm_i915_private *i915,
> >   			       u32 gt_iir)
> >   {
> >   	if (gt_iir & GT_RENDER_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> >   	if (gt_iir & GT_BSD_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
> >   	if (gt_iir & GT_BLT_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[BCS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[BCS0]);
> >   	if (gt_iir & (GT_BLT_CS_ERROR_INTERRUPT |
> >   		      GT_BSD_CS_ERROR_INTERRUPT |
> >   		      GT_RENDER_CS_MASTER_ERROR_INTERRUPT))
> >   		DRM_DEBUG("Command parser error, gt_iir 0x%08x\n", gt_iir);
> > -	if (gt_iir & GT_PARITY_ERROR(dev_priv))
> > -		ivybridge_parity_error_irq_handler(dev_priv, gt_iir);
> > +	if (gt_iir & GT_PARITY_ERROR(i915))
> > +		ivybridge_parity_error_irq_handler(i915, gt_iir);
> >   }
> >   static void
> > @@ -1763,7 +1763,7 @@ static bool i9xx_port_hotplug_long_detect(enum hpd_pin pin, u32 val)
> >    *
> >    * Note that the caller is expected to zero out the masks initially.
> >    */
> > -static void intel_get_hpd_pins(struct drm_i915_private *dev_priv,
> > +static void intel_get_hpd_pins(struct drm_i915_private *i915,
> >   			       u32 *pin_mask, u32 *long_mask,
> >   			       u32 hotplug_trigger, u32 dig_hotplug_reg,
> >   			       const u32 hpd[HPD_NUM_PINS],
> > @@ -1786,25 +1786,25 @@ static void intel_get_hpd_pins(struct drm_i915_private *dev_priv,
> >   }
> > -static void gmbus_irq_handler(struct drm_i915_private *dev_priv)
> > +static void gmbus_irq_handler(struct drm_i915_private *i915)
> >   {
> > -	wake_up_all(&dev_priv->gmbus_wait_queue);
> > +	wake_up_all(&i915->gmbus_wait_queue);
> >   }
> > -static void dp_aux_irq_handler(struct drm_i915_private *dev_priv)
> > +static void dp_aux_irq_handler(struct drm_i915_private *i915)
> >   {
> > -	wake_up_all(&dev_priv->gmbus_wait_queue);
> > +	wake_up_all(&i915->gmbus_wait_queue);
> >   }
> >   #if defined(CONFIG_DEBUG_FS)
> > -static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> > +static void display_pipe_crc_irq_handler(struct drm_i915_private *i915,
> >   					 enum pipe pipe,
> >   					 u32 crc0, u32 crc1,
> >   					 u32 crc2, u32 crc3,
> >   					 u32 crc4)
> >   {
> > -	struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe];
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe];
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	u32 crcs[5] = { crc0, crc1, crc2, crc3, crc4 };
> >   	trace_intel_pipe_crc(crtc, crcs);
> > @@ -1819,7 +1819,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> >   	 * don't trust that one either.
> >   	 */
> >   	if (pipe_crc->skipped <= 0 ||
> > -	    (INTEL_GEN(dev_priv) >= 8 && pipe_crc->skipped == 1)) {
> > +	    (INTEL_GEN(i915) >= 8 && pipe_crc->skipped == 1)) {
> >   		pipe_crc->skipped++;
> >   		spin_unlock(&pipe_crc->lock);
> >   		return;
> > @@ -1832,7 +1832,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> >   }
> >   #else
> >   static inline void
> > -display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> > +display_pipe_crc_irq_handler(struct drm_i915_private *i915,
> >   			     enum pipe pipe,
> >   			     u32 crc0, u32 crc1,
> >   			     u32 crc2, u32 crc3,
> > @@ -1840,18 +1840,18 @@ display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> >   #endif
> > -static void hsw_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> > +static void hsw_pipe_crc_irq_handler(struct drm_i915_private *i915,
> >   				     enum pipe pipe)
> >   {
> > -	display_pipe_crc_irq_handler(dev_priv, pipe,
> > +	display_pipe_crc_irq_handler(i915, pipe,
> >   				     I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
> >   				     0, 0, 0, 0);
> >   }
> > -static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ivb_pipe_crc_irq_handler(struct drm_i915_private *i915,
> >   				     enum pipe pipe)
> >   {
> > -	display_pipe_crc_irq_handler(dev_priv, pipe,
> > +	display_pipe_crc_irq_handler(i915, pipe,
> >   				     I915_READ(PIPE_CRC_RES_1_IVB(pipe)),
> >   				     I915_READ(PIPE_CRC_RES_2_IVB(pipe)),
> >   				     I915_READ(PIPE_CRC_RES_3_IVB(pipe)),
> > @@ -1859,22 +1859,22 @@ static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> >   				     I915_READ(PIPE_CRC_RES_5_IVB(pipe)));
> >   }
> > -static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *i915,
> >   				      enum pipe pipe)
> >   {
> >   	u32 res1, res2;
> > -	if (INTEL_GEN(dev_priv) >= 3)
> > +	if (INTEL_GEN(i915) >= 3)
> >   		res1 = I915_READ(PIPE_CRC_RES_RES1_I915(pipe));
> >   	else
> >   		res1 = 0;
> > -	if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> >   		res2 = I915_READ(PIPE_CRC_RES_RES2_G4X(pipe));
> >   	else
> >   		res2 = 0;
> > -	display_pipe_crc_irq_handler(dev_priv, pipe,
> > +	display_pipe_crc_irq_handler(i915, pipe,
> >   				     I915_READ(PIPE_CRC_RES_RED(pipe)),
> >   				     I915_READ(PIPE_CRC_RES_GREEN(pipe)),
> >   				     I915_READ(PIPE_CRC_RES_BLUE(pipe)),
> > @@ -1903,34 +1903,34 @@ static void gen11_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir)
> >   	schedule_work(&rps->work);
> >   }
> > -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir)
> > +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> > -	if (pm_iir & dev_priv->pm_rps_events) {
> > -		spin_lock(&dev_priv->irq_lock);
> > -		gen6_mask_pm_irq(dev_priv, pm_iir & dev_priv->pm_rps_events);
> > +	if (pm_iir & i915->pm_rps_events) {
> > +		spin_lock(&i915->irq_lock);
> > +		gen6_mask_pm_irq(i915, pm_iir & i915->pm_rps_events);
> >   		if (rps->interrupts_enabled) {
> > -			rps->pm_iir |= pm_iir & dev_priv->pm_rps_events;
> > +			rps->pm_iir |= pm_iir & i915->pm_rps_events;
> >   			schedule_work(&rps->work);
> >   		}
> > -		spin_unlock(&dev_priv->irq_lock);
> > +		spin_unlock(&i915->irq_lock);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		return;
> >   	if (pm_iir & PM_VEBOX_USER_INTERRUPT)
> > -		intel_engine_breadcrumbs_irq(dev_priv->engine[VECS0]);
> > +		intel_engine_breadcrumbs_irq(i915->engine[VECS0]);
> >   	if (pm_iir & PM_VEBOX_CS_ERROR_INTERRUPT)
> >   		DRM_DEBUG("Command parser error, pm_iir 0x%08x\n", pm_iir);
> >   }
> > -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 gt_iir)
> > +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 gt_iir)
> >   {
> >   	if (gt_iir & GEN9_GUC_TO_HOST_INT_EVENT)
> > -		intel_guc_to_host_event_handler(&dev_priv->guc);
> > +		intel_guc_to_host_event_handler(&i915->guc);
> >   }
> >   static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir)
> > @@ -1939,32 +1939,32 @@ static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir)
> >   		intel_guc_to_host_event_handler(&i915->guc);
> >   }
> > -static void i9xx_pipestat_irq_reset(struct drm_i915_private *dev_priv)
> > +static void i9xx_pipestat_irq_reset(struct drm_i915_private *i915)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		I915_WRITE(PIPESTAT(pipe),
> >   			   PIPESTAT_INT_STATUS_MASK |
> >   			   PIPE_FIFO_UNDERRUN_STATUS);
> > -		dev_priv->pipestat_irq_mask[pipe] = 0;
> > +		i915->pipestat_irq_mask[pipe] = 0;
> >   	}
> >   }
> > -static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> > +static void i9xx_pipestat_irq_ack(struct drm_i915_private *i915,
> >   				  u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> >   {
> >   	int pipe;
> > -	spin_lock(&dev_priv->irq_lock);
> > +	spin_lock(&i915->irq_lock);
> > -	if (!dev_priv->display_irqs_enabled) {
> > -		spin_unlock(&dev_priv->irq_lock);
> > +	if (!i915->display_irqs_enabled) {
> > +		spin_unlock(&i915->irq_lock);
> >   		return;
> >   	}
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		i915_reg_t reg;
> >   		u32 status_mask, enable_mask, iir_bit = 0;
> > @@ -1991,14 +1991,14 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> >   			break;
> >   		}
> >   		if (iir & iir_bit)
> > -			status_mask |= dev_priv->pipestat_irq_mask[pipe];
> > +			status_mask |= i915->pipestat_irq_mask[pipe];
> >   		if (!status_mask)
> >   			continue;
> >   		reg = PIPESTAT(pipe);
> >   		pipe_stats[pipe] = I915_READ(reg) & status_mask;
> > -		enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> > +		enable_mask = i915_pipestat_enable_mask(i915, pipe);
> >   		/*
> >   		 * Clear the PIPE*STAT regs before the IIR
> > @@ -2014,104 +2014,104 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv,
> >   			I915_WRITE(reg, enable_mask);
> >   		}
> >   	}
> > -	spin_unlock(&dev_priv->irq_lock);
> > +	spin_unlock(&i915->irq_lock);
> >   }
> > -static void i8xx_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i8xx_pipestat_irq_handler(struct drm_i915_private *i915,
> >   				      u16 iir, u32 pipe_stats[I915_MAX_PIPES])
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> > -			i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> > +			i9xx_pipe_crc_irq_handler(i915, pipe);
> >   		if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   	}
> >   }
> > -static void i915_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i915_pipestat_irq_handler(struct drm_i915_private *i915,
> >   				      u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> >   {
> >   	bool blc_event = false;
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS)
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
> >   			blc_event = true;
> >   		if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> > -			i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> > +			i9xx_pipe_crc_irq_handler(i915, pipe);
> >   		if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   	}
> >   	if (blc_event || (iir & I915_ASLE_INTERRUPT))
> > -		intel_opregion_asle_intr(dev_priv);
> > +		intel_opregion_asle_intr(i915);
> >   }
> > -static void i965_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i965_pipestat_irq_handler(struct drm_i915_private *i915,
> >   				      u32 iir, u32 pipe_stats[I915_MAX_PIPES])
> >   {
> >   	bool blc_event = false;
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS)
> >   			blc_event = true;
> >   		if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> > -			i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> > +			i9xx_pipe_crc_irq_handler(i915, pipe);
> >   		if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   	}
> >   	if (blc_event || (iir & I915_ASLE_INTERRUPT))
> > -		intel_opregion_asle_intr(dev_priv);
> > +		intel_opregion_asle_intr(i915);
> >   	if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   }
> > -static void valleyview_pipestat_irq_handler(struct drm_i915_private *dev_priv,
> > +static void valleyview_pipestat_irq_handler(struct drm_i915_private *i915,
> >   					    u32 pipe_stats[I915_MAX_PIPES])
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS)
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS)
> > -			i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> > +			i9xx_pipe_crc_irq_handler(i915, pipe);
> >   		if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS)
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   	}
> >   	if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   }
> > -static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv)
> > +static u32 i9xx_hpd_irq_ack(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_status = 0, hotplug_status_mask;
> >   	int i;
> > -	if (IS_G4X(dev_priv) ||
> > -	    IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_G4X(i915) ||
> > +	    IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		hotplug_status_mask = HOTPLUG_INT_STATUS_G4X |
> >   			DP_AUX_CHANNEL_MASK_INT_STATUS_G4X;
> >   	else
> > @@ -2143,35 +2143,35 @@ static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv)
> >   	return hotplug_status;
> >   }
> > -static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i9xx_hpd_irq_handler(struct drm_i915_private *i915,
> >   				 u32 hotplug_status)
> >   {
> >   	u32 pin_mask = 0, long_mask = 0;
> > -	if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> > -	    IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> > +	    IS_CHERRYVIEW(i915)) {
> >   		u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_G4X;
> >   		if (hotplug_trigger) {
> > -			intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +			intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   					   hotplug_trigger, hotplug_trigger,
> >   					   hpd_status_g4x,
> >   					   i9xx_port_hotplug_long_detect);
> > -			intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +			intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   		}
> >   		if (hotplug_status & DP_AUX_CHANNEL_MASK_INT_STATUS_G4X)
> > -			dp_aux_irq_handler(dev_priv);
> > +			dp_aux_irq_handler(i915);
> >   	} else {
> >   		u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_I915;
> >   		if (hotplug_trigger) {
> > -			intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +			intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   					   hotplug_trigger, hotplug_trigger,
> >   					   hpd_status_i915,
> >   					   i9xx_port_hotplug_long_detect);
> > -			intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +			intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   		}
> >   	}
> >   }
> > @@ -2179,14 +2179,14 @@ static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	do {
> >   		u32 iir, gt_iir, pm_iir;
> > @@ -2226,15 +2226,15 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> >   			I915_WRITE(GEN6_PMIIR, pm_iir);
> >   		if (iir & I915_DISPLAY_PORT_INTERRUPT)
> > -			hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> > +			hotplug_status = i9xx_hpd_irq_ack(i915);
> >   		/* Call regardless, as some status bits might not be
> >   		 * signalled in iir */
> > -		i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> > +		i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
> >   		if (iir & (I915_LPE_PIPE_A_INTERRUPT |
> >   			   I915_LPE_PIPE_B_INTERRUPT))
> > -			intel_lpe_audio_irq_handler(dev_priv);
> > +			intel_lpe_audio_irq_handler(i915);
> >   		/*
> >   		 * VLV_IIR is single buffered, and reflects the level
> > @@ -2247,17 +2247,17 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> >   		I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
> >   		if (gt_iir)
> > -			snb_gt_irq_handler(dev_priv, gt_iir);
> > +			snb_gt_irq_handler(i915, gt_iir);
> >   		if (pm_iir)
> > -			gen6_rps_irq_handler(dev_priv, pm_iir);
> > +			gen6_rps_irq_handler(i915, pm_iir);
> >   		if (hotplug_status)
> > -			i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> > +			i9xx_hpd_irq_handler(i915, hotplug_status);
> > -		valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
> > +		valleyview_pipestat_irq_handler(i915, pipe_stats);
> >   	} while (0);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> > @@ -2265,14 +2265,14 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg)
> >   static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	do {
> >   		u32 master_ctl, iir;
> > @@ -2306,19 +2306,19 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> >   		ier = I915_READ(VLV_IER);
> >   		I915_WRITE(VLV_IER, 0);
> > -		gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
> > +		gen8_gt_irq_ack(i915, master_ctl, gt_iir);
> >   		if (iir & I915_DISPLAY_PORT_INTERRUPT)
> > -			hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> > +			hotplug_status = i9xx_hpd_irq_ack(i915);
> >   		/* Call regardless, as some status bits might not be
> >   		 * signalled in iir */
> > -		i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> > +		i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
> >   		if (iir & (I915_LPE_PIPE_A_INTERRUPT |
> >   			   I915_LPE_PIPE_B_INTERRUPT |
> >   			   I915_LPE_PIPE_C_INTERRUPT))
> > -			intel_lpe_audio_irq_handler(dev_priv);
> > +			intel_lpe_audio_irq_handler(i915);
> >   		/*
> >   		 * VLV_IIR is single buffered, and reflects the level
> > @@ -2330,20 +2330,20 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg)
> >   		I915_WRITE(VLV_IER, ier);
> >   		I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
> > -		gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir);
> > +		gen8_gt_irq_handler(i915, master_ctl, gt_iir);
> >   		if (hotplug_status)
> > -			i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> > +			i9xx_hpd_irq_handler(i915, hotplug_status);
> > -		valleyview_pipestat_irq_handler(dev_priv, pipe_stats);
> > +		valleyview_pipestat_irq_handler(i915, pipe_stats);
> >   	} while (0);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> > -static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ibx_hpd_irq_handler(struct drm_i915_private *i915,
> >   				u32 hotplug_trigger,
> >   				const u32 hpd[HPD_NUM_PINS])
> >   {
> > @@ -2368,19 +2368,19 @@ static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	if (!hotplug_trigger)
> >   		return;
> > -	intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> > +	intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> >   			   dig_hotplug_reg, hpd,
> >   			   pch_port_hotplug_long_detect);
> > -	intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +	intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   }
> > -static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> > +static void ibx_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> >   {
> >   	int pipe;
> >   	u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK;
> > -	ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ibx);
> > +	ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_ibx);
> >   	if (pch_iir & SDE_AUDIO_POWER_MASK) {
> >   		int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK) >>
> > @@ -2390,10 +2390,10 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   	}
> >   	if (pch_iir & SDE_AUX_MASK)
> > -		dp_aux_irq_handler(dev_priv);
> > +		dp_aux_irq_handler(i915);
> >   	if (pch_iir & SDE_GMBUS)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   	if (pch_iir & SDE_AUDIO_HDCP_MASK)
> >   		DRM_DEBUG_DRIVER("PCH HDCP audio interrupt\n");
> > @@ -2405,7 +2405,7 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		DRM_ERROR("PCH poison interrupt\n");
> >   	if (pch_iir & SDE_FDI_MASK)
> > -		for_each_pipe(dev_priv, pipe)
> > +		for_each_pipe(i915, pipe)
> >   			DRM_DEBUG_DRIVER("  pipe %c FDI IIR: 0x%08x\n",
> >   					 pipe_name(pipe),
> >   					 I915_READ(FDI_RX_IIR(pipe)));
> > @@ -2417,13 +2417,13 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		DRM_DEBUG_DRIVER("PCH transcoder CRC error interrupt\n");
> >   	if (pch_iir & SDE_TRANSA_FIFO_UNDER)
> > -		intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_A);
> > +		intel_pch_fifo_underrun_irq_handler(i915, PIPE_A);
> >   	if (pch_iir & SDE_TRANSB_FIFO_UNDER)
> > -		intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_B);
> > +		intel_pch_fifo_underrun_irq_handler(i915, PIPE_B);
> >   }
> > -static void ivb_err_int_handler(struct drm_i915_private *dev_priv)
> > +static void ivb_err_int_handler(struct drm_i915_private *i915)
> >   {
> >   	u32 err_int = I915_READ(GEN7_ERR_INT);
> >   	enum pipe pipe;
> > @@ -2431,22 +2431,22 @@ static void ivb_err_int_handler(struct drm_i915_private *dev_priv)
> >   	if (err_int & ERR_INT_POISON)
> >   		DRM_ERROR("Poison interrupt\n");
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (err_int & ERR_INT_FIFO_UNDERRUN(pipe))
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   		if (err_int & ERR_INT_PIPE_CRC_DONE(pipe)) {
> > -			if (IS_IVYBRIDGE(dev_priv))
> > -				ivb_pipe_crc_irq_handler(dev_priv, pipe);
> > +			if (IS_IVYBRIDGE(i915))
> > +				ivb_pipe_crc_irq_handler(i915, pipe);
> >   			else
> > -				hsw_pipe_crc_irq_handler(dev_priv, pipe);
> > +				hsw_pipe_crc_irq_handler(i915, pipe);
> >   		}
> >   	}
> >   	I915_WRITE(GEN7_ERR_INT, err_int);
> >   }
> > -static void cpt_serr_int_handler(struct drm_i915_private *dev_priv)
> > +static void cpt_serr_int_handler(struct drm_i915_private *i915)
> >   {
> >   	u32 serr_int = I915_READ(SERR_INT);
> >   	enum pipe pipe;
> > @@ -2454,19 +2454,19 @@ static void cpt_serr_int_handler(struct drm_i915_private *dev_priv)
> >   	if (serr_int & SERR_INT_POISON)
> >   		DRM_ERROR("PCH poison interrupt\n");
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		if (serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pipe))
> > -			intel_pch_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_pch_fifo_underrun_irq_handler(i915, pipe);
> >   	I915_WRITE(SERR_INT, serr_int);
> >   }
> > -static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> > +static void cpt_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> >   {
> >   	int pipe;
> >   	u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_CPT;
> > -	ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_cpt);
> > +	ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_cpt);
> >   	if (pch_iir & SDE_AUDIO_POWER_MASK_CPT) {
> >   		int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK_CPT) >>
> > @@ -2476,10 +2476,10 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   	}
> >   	if (pch_iir & SDE_AUX_MASK_CPT)
> > -		dp_aux_irq_handler(dev_priv);
> > +		dp_aux_irq_handler(i915);
> >   	if (pch_iir & SDE_GMBUS_CPT)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   	if (pch_iir & SDE_AUDIO_CP_REQ_CPT)
> >   		DRM_DEBUG_DRIVER("Audio CP request interrupt\n");
> > @@ -2488,16 +2488,16 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		DRM_DEBUG_DRIVER("Audio CP change interrupt\n");
> >   	if (pch_iir & SDE_FDI_MASK_CPT)
> > -		for_each_pipe(dev_priv, pipe)
> > +		for_each_pipe(i915, pipe)
> >   			DRM_DEBUG_DRIVER("  pipe %c FDI IIR: 0x%08x\n",
> >   					 pipe_name(pipe),
> >   					 I915_READ(FDI_RX_IIR(pipe)));
> >   	if (pch_iir & SDE_ERROR_CPT)
> > -		cpt_serr_int_handler(dev_priv);
> > +		cpt_serr_int_handler(i915);
> >   }
> > -static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> > +static void icp_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> >   {
> >   	u32 ddi_hotplug_trigger = pch_iir & SDE_DDI_MASK_ICP;
> >   	u32 tc_hotplug_trigger = pch_iir & SDE_TC_MASK_ICP;
> > @@ -2509,7 +2509,7 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_DDI);
> >   		I915_WRITE(SHOTPLUG_CTL_DDI, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   				   ddi_hotplug_trigger,
> >   				   dig_hotplug_reg, hpd_icp,
> >   				   icp_ddi_port_hotplug_long_detect);
> > @@ -2521,20 +2521,20 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_TC);
> >   		I915_WRITE(SHOTPLUG_CTL_TC, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   				   tc_hotplug_trigger,
> >   				   dig_hotplug_reg, hpd_icp,
> >   				   icp_tc_port_hotplug_long_detect);
> >   	}
> >   	if (pin_mask)
> > -		intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +		intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   	if (pch_iir & SDE_GMBUS_ICP)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   }
> > -static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> > +static void spt_irq_handler(struct drm_i915_private *i915, u32 pch_iir)
> >   {
> >   	u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_SPT &
> >   		~SDE_PORTE_HOTPLUG_SPT;
> > @@ -2547,7 +2547,7 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
> >   		I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   				   hotplug_trigger, dig_hotplug_reg, hpd_spt,
> >   				   spt_port_hotplug_long_detect);
> >   	}
> > @@ -2558,19 +2558,19 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir)
> >   		dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG2);
> >   		I915_WRITE(PCH_PORT_HOTPLUG2, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask,
> >   				   hotplug2_trigger, dig_hotplug_reg, hpd_spt,
> >   				   spt_port_hotplug2_long_detect);
> >   	}
> >   	if (pin_mask)
> > -		intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +		intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   	if (pch_iir & SDE_GMBUS_CPT)
> > -		gmbus_irq_handler(dev_priv);
> > +		gmbus_irq_handler(i915);
> >   }
> > -static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ilk_hpd_irq_handler(struct drm_i915_private *i915,
> >   				u32 hotplug_trigger,
> >   				const u32 hpd[HPD_NUM_PINS])
> >   {
> > @@ -2579,94 +2579,94 @@ static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	dig_hotplug_reg = I915_READ(DIGITAL_PORT_HOTPLUG_CNTRL);
> >   	I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, dig_hotplug_reg);
> > -	intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> > +	intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> >   			   dig_hotplug_reg, hpd,
> >   			   ilk_port_hotplug_long_detect);
> > -	intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +	intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   }
> > -static void ilk_display_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ilk_display_irq_handler(struct drm_i915_private *i915,
> >   				    u32 de_iir)
> >   {
> >   	enum pipe pipe;
> >   	u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG;
> >   	if (hotplug_trigger)
> > -		ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ilk);
> > +		ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ilk);
> >   	if (de_iir & DE_AUX_CHANNEL_A)
> > -		dp_aux_irq_handler(dev_priv);
> > +		dp_aux_irq_handler(i915);
> >   	if (de_iir & DE_GSE)
> > -		intel_opregion_asle_intr(dev_priv);
> > +		intel_opregion_asle_intr(i915);
> >   	if (de_iir & DE_POISON)
> >   		DRM_ERROR("Poison interrupt\n");
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (de_iir & DE_PIPE_VBLANK(pipe))
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (de_iir & DE_PIPE_FIFO_UNDERRUN(pipe))
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   		if (de_iir & DE_PIPE_CRC_DONE(pipe))
> > -			i9xx_pipe_crc_irq_handler(dev_priv, pipe);
> > +			i9xx_pipe_crc_irq_handler(i915, pipe);
> >   	}
> >   	/* check event from PCH */
> >   	if (de_iir & DE_PCH_EVENT) {
> >   		u32 pch_iir = I915_READ(SDEIIR);
> > -		if (HAS_PCH_CPT(dev_priv))
> > -			cpt_irq_handler(dev_priv, pch_iir);
> > +		if (HAS_PCH_CPT(i915))
> > +			cpt_irq_handler(i915, pch_iir);
> >   		else
> > -			ibx_irq_handler(dev_priv, pch_iir);
> > +			ibx_irq_handler(i915, pch_iir);
> >   		/* should clear PCH hotplug event before clear CPU irq */
> >   		I915_WRITE(SDEIIR, pch_iir);
> >   	}
> > -	if (IS_GEN(dev_priv, 5) && de_iir & DE_PCU_EVENT)
> > -		ironlake_rps_change_irq_handler(dev_priv);
> > +	if (IS_GEN(i915, 5) && de_iir & DE_PCU_EVENT)
> > +		ironlake_rps_change_irq_handler(i915);
> >   }
> > -static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
> > +static void ivb_display_irq_handler(struct drm_i915_private *i915,
> >   				    u32 de_iir)
> >   {
> >   	enum pipe pipe;
> >   	u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG_IVB;
> >   	if (hotplug_trigger)
> > -		ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ivb);
> > +		ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ivb);
> >   	if (de_iir & DE_ERR_INT_IVB)
> > -		ivb_err_int_handler(dev_priv);
> > +		ivb_err_int_handler(i915);
> >   	if (de_iir & DE_EDP_PSR_INT_HSW) {
> >   		u32 psr_iir = I915_READ(EDP_PSR_IIR);
> > -		intel_psr_irq_handler(dev_priv, psr_iir);
> > +		intel_psr_irq_handler(i915, psr_iir);
> >   		I915_WRITE(EDP_PSR_IIR, psr_iir);
> >   	}
> >   	if (de_iir & DE_AUX_CHANNEL_A_IVB)
> > -		dp_aux_irq_handler(dev_priv);
> > +		dp_aux_irq_handler(i915);
> >   	if (de_iir & DE_GSE_IVB)
> > -		intel_opregion_asle_intr(dev_priv);
> > +		intel_opregion_asle_intr(i915);
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (de_iir & (DE_PIPE_VBLANK_IVB(pipe)))
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   	}
> >   	/* check event from PCH */
> > -	if (!HAS_PCH_NOP(dev_priv) && (de_iir & DE_PCH_EVENT_IVB)) {
> > +	if (!HAS_PCH_NOP(i915) && (de_iir & DE_PCH_EVENT_IVB)) {
> >   		u32 pch_iir = I915_READ(SDEIIR);
> > -		cpt_irq_handler(dev_priv, pch_iir);
> > +		cpt_irq_handler(i915, pch_iir);
> >   		/* clear PCH hotplug event before clear CPU irq */
> >   		I915_WRITE(SDEIIR, pch_iir);
> > @@ -2684,15 +2684,15 @@ static void ivb_display_irq_handler(struct drm_i915_private *dev_priv,
> >   static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 de_iir, gt_iir, de_ier, sde_ier = 0;
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	/* disable master interrupt before clearing iir  */
> >   	de_ier = I915_READ(DEIER);
> > @@ -2703,7 +2703,7 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> >   	 * able to process them after we restore SDEIER (as soon as we restore
> >   	 * it, we'll get an interrupt if SDEIIR still has something to process
> >   	 * due to its back queue). */
> > -	if (!HAS_PCH_NOP(dev_priv)) {
> > +	if (!HAS_PCH_NOP(i915)) {
> >   		sde_ier = I915_READ(SDEIER);
> >   		I915_WRITE(SDEIER, 0);
> >   	}
> > @@ -2714,42 +2714,42 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg)
> >   	if (gt_iir) {
> >   		I915_WRITE(GTIIR, gt_iir);
> >   		ret = IRQ_HANDLED;
> > -		if (INTEL_GEN(dev_priv) >= 6)
> > -			snb_gt_irq_handler(dev_priv, gt_iir);
> > +		if (INTEL_GEN(i915) >= 6)
> > +			snb_gt_irq_handler(i915, gt_iir);
> >   		else
> > -			ilk_gt_irq_handler(dev_priv, gt_iir);
> > +			ilk_gt_irq_handler(i915, gt_iir);
> >   	}
> >   	de_iir = I915_READ(DEIIR);
> >   	if (de_iir) {
> >   		I915_WRITE(DEIIR, de_iir);
> >   		ret = IRQ_HANDLED;
> > -		if (INTEL_GEN(dev_priv) >= 7)
> > -			ivb_display_irq_handler(dev_priv, de_iir);
> > +		if (INTEL_GEN(i915) >= 7)
> > +			ivb_display_irq_handler(i915, de_iir);
> >   		else
> > -			ilk_display_irq_handler(dev_priv, de_iir);
> > +			ilk_display_irq_handler(i915, de_iir);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 6) {
> > +	if (INTEL_GEN(i915) >= 6) {
> >   		u32 pm_iir = I915_READ(GEN6_PMIIR);
> >   		if (pm_iir) {
> >   			I915_WRITE(GEN6_PMIIR, pm_iir);
> >   			ret = IRQ_HANDLED;
> > -			gen6_rps_irq_handler(dev_priv, pm_iir);
> > +			gen6_rps_irq_handler(i915, pm_iir);
> >   		}
> >   	}
> >   	I915_WRITE(DEIER, de_ier);
> > -	if (!HAS_PCH_NOP(dev_priv))
> > +	if (!HAS_PCH_NOP(i915))
> >   		I915_WRITE(SDEIER, sde_ier);
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> > -static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +static void bxt_hpd_irq_handler(struct drm_i915_private *i915,
> >   				u32 hotplug_trigger,
> >   				const u32 hpd[HPD_NUM_PINS])
> >   {
> > @@ -2758,14 +2758,14 @@ static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG);
> >   	I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg);
> > -	intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger,
> > +	intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger,
> >   			   dig_hotplug_reg, hpd,
> >   			   bxt_port_hotplug_long_detect);
> > -	intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +	intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   }
> > -static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> > +static void gen11_hpd_irq_handler(struct drm_i915_private *i915, u32 iir)
> >   {
> >   	u32 pin_mask = 0, long_mask = 0;
> >   	u32 trigger_tc = iir & GEN11_DE_TC_HOTPLUG_MASK;
> > @@ -2777,7 +2777,7 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> >   		dig_hotplug_reg = I915_READ(GEN11_TC_HOTPLUG_CTL);
> >   		I915_WRITE(GEN11_TC_HOTPLUG_CTL, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tc,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tc,
> >   				   dig_hotplug_reg, hpd_gen11,
> >   				   gen11_port_hotplug_long_detect);
> >   	}
> > @@ -2788,30 +2788,30 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir)
> >   		dig_hotplug_reg = I915_READ(GEN11_TBT_HOTPLUG_CTL);
> >   		I915_WRITE(GEN11_TBT_HOTPLUG_CTL, dig_hotplug_reg);
> > -		intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tbt,
> > +		intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tbt,
> >   				   dig_hotplug_reg, hpd_gen11,
> >   				   gen11_port_hotplug_long_detect);
> >   	}
> >   	if (pin_mask)
> > -		intel_hpd_irq_handler(dev_priv, pin_mask, long_mask);
> > +		intel_hpd_irq_handler(i915, pin_mask, long_mask);
> >   	else
> >   		DRM_ERROR("Unexpected DE HPD interrupt 0x%08x\n", iir);
> >   }
> > -static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv)
> > +static u32 gen8_de_port_aux_mask(struct drm_i915_private *i915)
> >   {
> >   	u32 mask = GEN8_AUX_CHANNEL_A;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		mask |= GEN9_AUX_CHANNEL_B |
> >   			GEN9_AUX_CHANNEL_C |
> >   			GEN9_AUX_CHANNEL_D;
> > -	if (IS_CNL_WITH_PORT_F(dev_priv))
> > +	if (IS_CNL_WITH_PORT_F(i915))
> >   		mask |= CNL_AUX_CHANNEL_F;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		mask |= ICL_AUX_CHANNEL_E |
> >   			CNL_AUX_CHANNEL_F;
> > @@ -2819,7 +2819,7 @@ static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv)
> >   }
> >   static irqreturn_t
> > -gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> > +gen8_de_irq_handler(struct drm_i915_private *i915, u32 master_ctl)
> >   {
> >   	irqreturn_t ret = IRQ_NONE;
> >   	u32 iir;
> > @@ -2834,14 +2834,14 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   			ret = IRQ_HANDLED;
> >   			if (iir & GEN8_DE_MISC_GSE) {
> > -				intel_opregion_asle_intr(dev_priv);
> > +				intel_opregion_asle_intr(i915);
> >   				found = true;
> >   			}
> >   			if (iir & GEN8_DE_EDP_PSR) {
> >   				u32 psr_iir = I915_READ(EDP_PSR_IIR);
> > -				intel_psr_irq_handler(dev_priv, psr_iir);
> > +				intel_psr_irq_handler(i915, psr_iir);
> >   				I915_WRITE(EDP_PSR_IIR, psr_iir);
> >   				found = true;
> >   			}
> > @@ -2853,12 +2853,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   			DRM_ERROR("The master control interrupt lied (DE MISC)!\n");
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) {
> > +	if (INTEL_GEN(i915) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) {
> >   		iir = I915_READ(GEN11_DE_HPD_IIR);
> >   		if (iir) {
> >   			I915_WRITE(GEN11_DE_HPD_IIR, iir);
> >   			ret = IRQ_HANDLED;
> > -			gen11_hpd_irq_handler(dev_priv, iir);
> > +			gen11_hpd_irq_handler(i915, iir);
> >   		} else {
> >   			DRM_ERROR("The master control interrupt lied, (DE HPD)!\n");
> >   		}
> > @@ -2873,29 +2873,29 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   			I915_WRITE(GEN8_DE_PORT_IIR, iir);
> >   			ret = IRQ_HANDLED;
> > -			if (iir & gen8_de_port_aux_mask(dev_priv)) {
> > -				dp_aux_irq_handler(dev_priv);
> > +			if (iir & gen8_de_port_aux_mask(i915)) {
> > +				dp_aux_irq_handler(i915);
> >   				found = true;
> >   			}
> > -			if (IS_GEN9_LP(dev_priv)) {
> > +			if (IS_GEN9_LP(i915)) {
> >   				tmp_mask = iir & BXT_DE_PORT_HOTPLUG_MASK;
> >   				if (tmp_mask) {
> > -					bxt_hpd_irq_handler(dev_priv, tmp_mask,
> > +					bxt_hpd_irq_handler(i915, tmp_mask,
> >   							    hpd_bxt);
> >   					found = true;
> >   				}
> > -			} else if (IS_BROADWELL(dev_priv)) {
> > +			} else if (IS_BROADWELL(i915)) {
> >   				tmp_mask = iir & GEN8_PORT_DP_A_HOTPLUG;
> >   				if (tmp_mask) {
> > -					ilk_hpd_irq_handler(dev_priv,
> > +					ilk_hpd_irq_handler(i915,
> >   							    tmp_mask, hpd_bdw);
> >   					found = true;
> >   				}
> >   			}
> > -			if (IS_GEN9_LP(dev_priv) && (iir & BXT_DE_PORT_GMBUS)) {
> > -				gmbus_irq_handler(dev_priv);
> > +			if (IS_GEN9_LP(i915) && (iir & BXT_DE_PORT_GMBUS)) {
> > +				gmbus_irq_handler(i915);
> >   				found = true;
> >   			}
> > @@ -2906,7 +2906,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   			DRM_ERROR("The master control interrupt lied (DE PORT)!\n");
> >   	}
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		u32 fault_errors;
> >   		if (!(master_ctl & GEN8_DE_PIPE_IRQ(pipe)))
> > @@ -2922,16 +2922,16 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   		I915_WRITE(GEN8_DE_PIPE_IIR(pipe), iir);
> >   		if (iir & GEN8_PIPE_VBLANK)
> > -			drm_handle_vblank(&dev_priv->drm, pipe);
> > +			drm_handle_vblank(&i915->drm, pipe);
> >   		if (iir & GEN8_PIPE_CDCLK_CRC_DONE)
> > -			hsw_pipe_crc_irq_handler(dev_priv, pipe);
> > +			hsw_pipe_crc_irq_handler(i915, pipe);
> >   		if (iir & GEN8_PIPE_FIFO_UNDERRUN)
> > -			intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe);
> > +			intel_cpu_fifo_underrun_irq_handler(i915, pipe);
> >   		fault_errors = iir;
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			fault_errors &= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
> >   		else
> >   			fault_errors &= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
> > @@ -2942,7 +2942,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   				  fault_errors);
> >   	}
> > -	if (HAS_PCH_SPLIT(dev_priv) && !HAS_PCH_NOP(dev_priv) &&
> > +	if (HAS_PCH_SPLIT(i915) && !HAS_PCH_NOP(i915) &&
> >   	    master_ctl & GEN8_DE_PCH_IRQ) {
> >   		/*
> >   		 * FIXME(BDW): Assume for now that the new interrupt handling
> > @@ -2954,12 +2954,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl)
> >   			I915_WRITE(SDEIIR, iir);
> >   			ret = IRQ_HANDLED;
> > -			if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> > -				icp_irq_handler(dev_priv, iir);
> > -			else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT)
> > -				spt_irq_handler(dev_priv, iir);
> > +			if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> > +				icp_irq_handler(i915, iir);
> > +			else if (INTEL_PCH_TYPE(i915) >= PCH_SPT)
> > +				spt_irq_handler(i915, iir);
> >   			else
> > -				cpt_irq_handler(dev_priv, iir);
> > +				cpt_irq_handler(i915, iir);
> >   		} else {
> >   			/*
> >   			 * Like on previous PCH there seems to be something
> > @@ -2992,12 +2992,12 @@ static inline void gen8_master_intr_enable(void __iomem * const regs)
> >   static irqreturn_t gen8_irq_handler(int irq, void *arg)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(arg);
> > -	void __iomem * const regs = dev_priv->uncore.regs;
> > +	struct drm_i915_private *i915 = to_i915(arg);
> > +	void __iomem * const regs = i915->uncore.regs;
> >   	u32 master_ctl;
> >   	u32 gt_iir[4];
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	master_ctl = gen8_master_intr_disable(regs);
> > @@ -3007,18 +3007,18 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg)
> >   	}
> >   	/* Find, clear, then process each source of interrupt */
> > -	gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir);
> > +	gen8_gt_irq_ack(i915, master_ctl, gt_iir);
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> >   	if (master_ctl & ~GEN8_GT_IRQS) {
> > -		disable_rpm_wakeref_asserts(dev_priv);
> > -		gen8_de_irq_handler(dev_priv, master_ctl);
> > -		enable_rpm_wakeref_asserts(dev_priv);
> > +		disable_rpm_wakeref_asserts(i915);
> > +		gen8_de_irq_handler(i915, master_ctl);
> > +		enable_rpm_wakeref_asserts(i915);
> >   	}
> >   	gen8_master_intr_enable(regs);
> > -	gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir);
> > +	gen8_gt_irq_handler(i915, master_ctl, gt_iir);
> >   	return IRQ_HANDLED;
> >   }
> > @@ -3149,9 +3149,9 @@ gen11_gt_irq_handler(struct drm_i915_private * const i915,
> >   }
> >   static u32
> > -gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl)
> > +gen11_gu_misc_irq_ack(struct drm_i915_private *i915, const u32 master_ctl)
> >   {
> > -	void __iomem * const regs = dev_priv->uncore.regs;
> > +	void __iomem * const regs = i915->uncore.regs;
> >   	u32 iir;
> >   	if (!(master_ctl & GEN11_GU_MISC_IRQ))
> > @@ -3165,10 +3165,10 @@ gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl)
> >   }
> >   static void
> > -gen11_gu_misc_irq_handler(struct drm_i915_private *dev_priv, const u32 iir)
> > +gen11_gu_misc_irq_handler(struct drm_i915_private *i915, const u32 iir)
> >   {
> >   	if (iir & GEN11_GU_MISC_GSE)
> > -		intel_opregion_asle_intr(dev_priv);
> > +		intel_opregion_asle_intr(i915);
> >   }
> >   static inline u32 gen11_master_intr_disable(void __iomem * const regs)
> > @@ -3235,54 +3235,54 @@ static irqreturn_t gen11_irq_handler(int irq, void *arg)
> >    */
> >   static int i8xx_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	i915_enable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	i915_enable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   	return 0;
> >   }
> >   static int i945gm_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	if (dev_priv->i945gm_vblank.enabled++ == 0)
> > -		schedule_work(&dev_priv->i945gm_vblank.work);
> > +	if (i915->i945gm_vblank.enabled++ == 0)
> > +		schedule_work(&i915->i945gm_vblank.work);
> >   	return i8xx_enable_vblank(dev, pipe);
> >   }
> >   static int i965_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	i915_enable_pipestat(dev_priv, pipe,
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	i915_enable_pipestat(i915, pipe,
> >   			     PIPE_START_VBLANK_INTERRUPT_STATUS);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   	return 0;
> >   }
> >   static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	u32 bit = INTEL_GEN(dev_priv) >= 7 ?
> > +	u32 bit = INTEL_GEN(i915) >= 7 ?
> >   		DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	ilk_enable_display_irq(dev_priv, bit);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	ilk_enable_display_irq(i915, bit);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   	/* Even though there is no DMC, frame counter can get stuck when
> >   	 * PSR is active as no frames are generated.
> >   	 */
> > -	if (HAS_PSR(dev_priv))
> > +	if (HAS_PSR(i915))
> >   		drm_vblank_restore(dev, pipe);
> >   	return 0;
> > @@ -3290,17 +3290,17 @@ static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   	/* Even if there is no DMC, frame counter can get stuck when
> >   	 * PSR is active as no frames are generated, so check only for PSR.
> >   	 */
> > -	if (HAS_PSR(dev_priv))
> > +	if (HAS_PSR(i915))
> >   		drm_vblank_restore(dev, pipe);
> >   	return 0;
> > @@ -3311,60 +3311,60 @@ static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe)
> >    */
> >   static void i8xx_disable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	i915_disable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	i915_disable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   }
> >   static void i945gm_disable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	i8xx_disable_vblank(dev, pipe);
> > -	if (--dev_priv->i945gm_vblank.enabled == 0)
> > -		schedule_work(&dev_priv->i945gm_vblank.work);
> > +	if (--i915->i945gm_vblank.enabled == 0)
> > +		schedule_work(&i915->i945gm_vblank.work);
> >   }
> >   static void i965_disable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	i915_disable_pipestat(dev_priv, pipe,
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	i915_disable_pipestat(i915, pipe,
> >   			      PIPE_START_VBLANK_INTERRUPT_STATUS);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   }
> >   static void ironlake_disable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	u32 bit = INTEL_GEN(dev_priv) >= 7 ?
> > +	u32 bit = INTEL_GEN(i915) >= 7 ?
> >   		DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe);
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	ilk_disable_display_irq(dev_priv, bit);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	ilk_disable_display_irq(i915, bit);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   }
> >   static void gen8_disable_vblank(struct drm_device *dev, unsigned int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, irqflags);
> > -	bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags);
> > +	spin_lock_irqsave(&i915->irq_lock, irqflags);
> > +	bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK);
> > +	spin_unlock_irqrestore(&i915->irq_lock, irqflags);
> >   }
> >   static void i945gm_vblank_work_func(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private, i945gm_vblank.work);
> >   	/*
> > @@ -3372,9 +3372,9 @@ static void i945gm_vblank_work_func(struct work_struct *work)
> >   	 * hence we want to prevent C3 usage while vblank interrupts
> >   	 * are enabled.
> >   	 */
> > -	pm_qos_update_request(&dev_priv->i945gm_vblank.pm_qos,
> > -			      READ_ONCE(dev_priv->i945gm_vblank.enabled) ?
> > -			      dev_priv->i945gm_vblank.c3_disable_latency :
> > +	pm_qos_update_request(&i915->i945gm_vblank.pm_qos,
> > +			      READ_ONCE(i915->i945gm_vblank.enabled) ?
> > +			      i915->i945gm_vblank.c3_disable_latency :
> >   			      PM_QOS_DEFAULT_VALUE);
> >   }
> > @@ -3398,34 +3398,34 @@ static int cstate_disable_latency(const char *name)
> >   	return 0;
> >   }
> > -static void i945gm_vblank_work_init(struct drm_i915_private *dev_priv)
> > +static void i945gm_vblank_work_init(struct drm_i915_private *i915)
> >   {
> > -	INIT_WORK(&dev_priv->i945gm_vblank.work,
> > +	INIT_WORK(&i915->i945gm_vblank.work,
> >   		  i945gm_vblank_work_func);
> > -	dev_priv->i945gm_vblank.c3_disable_latency =
> > +	i915->i945gm_vblank.c3_disable_latency =
> >   		cstate_disable_latency("C3");
> > -	pm_qos_add_request(&dev_priv->i945gm_vblank.pm_qos,
> > +	pm_qos_add_request(&i915->i945gm_vblank.pm_qos,
> >   			   PM_QOS_CPU_DMA_LATENCY,
> >   			   PM_QOS_DEFAULT_VALUE);
> >   }
> > -static void i945gm_vblank_work_fini(struct drm_i915_private *dev_priv)
> > +static void i945gm_vblank_work_fini(struct drm_i915_private *i915)
> >   {
> > -	cancel_work_sync(&dev_priv->i945gm_vblank.work);
> > -	pm_qos_remove_request(&dev_priv->i945gm_vblank.pm_qos);
> > +	cancel_work_sync(&i915->i945gm_vblank.work);
> > +	pm_qos_remove_request(&i915->i945gm_vblank.pm_qos);
> >   }
> > -static void ibx_irq_reset(struct drm_i915_private *dev_priv)
> > +static void ibx_irq_reset(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	if (HAS_PCH_NOP(dev_priv))
> > +	if (HAS_PCH_NOP(i915))
> >   		return;
> >   	GEN3_IRQ_RESET(uncore, SDE);
> > -	if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915))
> >   		I915_WRITE(SERR_INT, 0xffffffff);
> >   }
> > @@ -3439,9 +3439,9 @@ static void ibx_irq_reset(struct drm_i915_private *dev_priv)
> >    */
> >   static void ibx_irq_pre_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	if (HAS_PCH_NOP(dev_priv))
> > +	if (HAS_PCH_NOP(i915))
> >   		return;
> >   	WARN_ON(I915_READ(SDEIER) != 0);
> > @@ -3449,36 +3449,36 @@ static void ibx_irq_pre_postinstall(struct drm_device *dev)
> >   	POSTING_READ(SDEIER);
> >   }
> > -static void gen5_gt_irq_reset(struct drm_i915_private *dev_priv)
> > +static void gen5_gt_irq_reset(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	GEN3_IRQ_RESET(uncore, GT);
> > -	if (INTEL_GEN(dev_priv) >= 6)
> > +	if (INTEL_GEN(i915) >= 6)
> >   		GEN3_IRQ_RESET(uncore, GEN6_PM);
> >   }
> > -static void vlv_display_irq_reset(struct drm_i915_private *dev_priv)
> > +static void vlv_display_irq_reset(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK_CHV);
> >   	else
> >   		I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK);
> > -	i915_hotplug_interrupt_update_locked(dev_priv, 0xffffffff, 0);
> > +	i915_hotplug_interrupt_update_locked(i915, 0xffffffff, 0);
> >   	I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
> > -	i9xx_pipestat_irq_reset(dev_priv);
> > +	i9xx_pipestat_irq_reset(i915);
> >   	GEN3_IRQ_RESET(uncore, VLV_);
> > -	dev_priv->irq_mask = ~0u;
> > +	i915->irq_mask = ~0u;
> >   }
> > -static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
> > +static void vlv_display_irq_postinstall(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 pipestat_mask;
> >   	u32 enable_mask;
> > @@ -3486,9 +3486,9 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
> >   	pipestat_mask = PIPE_CRC_DONE_INTERRUPT_STATUS;
> > -	i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> > -	for_each_pipe(dev_priv, pipe)
> > -		i915_enable_pipestat(dev_priv, pipe, pipestat_mask);
> > +	i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> > +	for_each_pipe(i915, pipe)
> > +		i915_enable_pipestat(i915, pipe, pipestat_mask);
> >   	enable_mask = I915_DISPLAY_PORT_INTERRUPT |
> >   		I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> > @@ -3496,56 +3496,56 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv)
> >   		I915_LPE_PIPE_A_INTERRUPT |
> >   		I915_LPE_PIPE_B_INTERRUPT;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		enable_mask |= I915_DISPLAY_PIPE_C_EVENT_INTERRUPT |
> >   			I915_LPE_PIPE_C_INTERRUPT;
> > -	WARN_ON(dev_priv->irq_mask != ~0u);
> > +	WARN_ON(i915->irq_mask != ~0u);
> > -	dev_priv->irq_mask = ~enable_mask;
> > +	i915->irq_mask = ~enable_mask;
> > -	GEN3_IRQ_INIT(uncore, VLV_, dev_priv->irq_mask, enable_mask);
> > +	GEN3_IRQ_INIT(uncore, VLV_, i915->irq_mask, enable_mask);
> >   }
> >   /* drm_dma.h hooks
> >   */
> >   static void ironlake_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	GEN3_IRQ_RESET(uncore, DE);
> > -	if (IS_GEN(dev_priv, 7))
> > +	if (IS_GEN(i915, 7))
> >   		I915_WRITE(GEN7_ERR_INT, 0xffffffff);
> > -	if (IS_HASWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915)) {
> >   		I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> >   		I915_WRITE(EDP_PSR_IIR, 0xffffffff);
> >   	}
> > -	gen5_gt_irq_reset(dev_priv);
> > +	gen5_gt_irq_reset(i915);
> > -	ibx_irq_reset(dev_priv);
> > +	ibx_irq_reset(i915);
> >   }
> >   static void valleyview_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	I915_WRITE(VLV_MASTER_IER, 0);
> >   	POSTING_READ(VLV_MASTER_IER);
> > -	gen5_gt_irq_reset(dev_priv);
> > +	gen5_gt_irq_reset(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->display_irqs_enabled)
> > -		vlv_display_irq_reset(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->display_irqs_enabled)
> > +		vlv_display_irq_reset(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv)
> > +static void gen8_gt_irq_reset(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	GEN8_IRQ_RESET_NDX(uncore, GT, 0);
> >   	GEN8_IRQ_RESET_NDX(uncore, GT, 1);
> > @@ -3555,19 +3555,19 @@ static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv)
> >   static void gen8_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	int pipe;
> > -	gen8_master_intr_disable(dev_priv->uncore.regs);
> > +	gen8_master_intr_disable(i915->uncore.regs);
> > -	gen8_gt_irq_reset(dev_priv);
> > +	gen8_gt_irq_reset(i915);
> >   	I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> >   	I915_WRITE(EDP_PSR_IIR, 0xffffffff);
> > -	for_each_pipe(dev_priv, pipe)
> > -		if (intel_display_power_is_enabled(dev_priv,
> > +	for_each_pipe(i915, pipe)
> > +		if (intel_display_power_is_enabled(i915,
> >   						   POWER_DOMAIN_PIPE(pipe)))
> >   			GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
> > @@ -3575,11 +3575,11 @@ static void gen8_irq_reset(struct drm_device *dev)
> >   	GEN3_IRQ_RESET(uncore, GEN8_DE_MISC_);
> >   	GEN3_IRQ_RESET(uncore, GEN8_PCU_);
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > -		ibx_irq_reset(dev_priv);
> > +	if (HAS_PCH_SPLIT(i915))
> > +		ibx_irq_reset(i915);
> >   }
> > -static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv)
> > +static void gen11_gt_irq_reset(struct drm_i915_private *i915)
> >   {
> >   	/* Disable RCS, BCS, VCS and VECS class engines. */
> >   	I915_WRITE(GEN11_RENDER_COPY_INTR_ENABLE, 0);
> > @@ -3600,21 +3600,21 @@ static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv)
> >   static void gen11_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = dev->dev_private;
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = dev->dev_private;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	int pipe;
> > -	gen11_master_intr_disable(dev_priv->uncore.regs);
> > +	gen11_master_intr_disable(i915->uncore.regs);
> > -	gen11_gt_irq_reset(dev_priv);
> > +	gen11_gt_irq_reset(i915);
> >   	I915_WRITE(GEN11_DISPLAY_INT_CTL, 0);
> >   	I915_WRITE(EDP_PSR_IMR, 0xffffffff);
> >   	I915_WRITE(EDP_PSR_IIR, 0xffffffff);
> > -	for_each_pipe(dev_priv, pipe)
> > -		if (intel_display_power_is_enabled(dev_priv,
> > +	for_each_pipe(i915, pipe)
> > +		if (intel_display_power_is_enabled(i915,
> >   						   POWER_DOMAIN_PIPE(pipe)))
> >   			GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
> > @@ -3624,87 +3624,87 @@ static void gen11_irq_reset(struct drm_device *dev)
> >   	GEN3_IRQ_RESET(uncore, GEN11_GU_MISC_);
> >   	GEN3_IRQ_RESET(uncore, GEN8_PCU_);
> > -	if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> > +	if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> >   		GEN3_IRQ_RESET(uncore, SDE);
> >   }
> > -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
> > +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915,
> >   				     u8 pipe_mask)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 extra_ier = GEN8_PIPE_VBLANK | GEN8_PIPE_FIFO_UNDERRUN;
> >   	enum pipe pipe;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	if (!intel_irqs_enabled(dev_priv)) {
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +	if (!intel_irqs_enabled(i915)) {
> > +		spin_unlock_irq(&i915->irq_lock);
> >   		return;
> >   	}
> > -	for_each_pipe_masked(dev_priv, pipe, pipe_mask)
> > +	for_each_pipe_masked(i915, pipe, pipe_mask)
> >   		GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe,
> > -				  dev_priv->de_irq_mask[pipe],
> > -				  ~dev_priv->de_irq_mask[pipe] | extra_ier);
> > +				  i915->de_irq_mask[pipe],
> > +				  ~i915->de_irq_mask[pipe] | extra_ier);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv,
> > +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915,
> >   				     u8 pipe_mask)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	enum pipe pipe;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	if (!intel_irqs_enabled(dev_priv)) {
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +	if (!intel_irqs_enabled(i915)) {
> > +		spin_unlock_irq(&i915->irq_lock);
> >   		return;
> >   	}
> > -	for_each_pipe_masked(dev_priv, pipe, pipe_mask)
> > +	for_each_pipe_masked(i915, pipe, pipe_mask)
> >   		GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/* make sure we're done processing display irqs */
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	synchronize_irq(i915->drm.irq);
> >   }
> >   static void cherryview_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	I915_WRITE(GEN8_MASTER_IRQ, 0);
> >   	POSTING_READ(GEN8_MASTER_IRQ);
> > -	gen8_gt_irq_reset(dev_priv);
> > +	gen8_gt_irq_reset(i915);
> >   	GEN3_IRQ_RESET(uncore, GEN8_PCU_);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->display_irqs_enabled)
> > -		vlv_display_irq_reset(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->display_irqs_enabled)
> > +		vlv_display_irq_reset(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > -static u32 intel_hpd_enabled_irqs(struct drm_i915_private *dev_priv,
> > +static u32 intel_hpd_enabled_irqs(struct drm_i915_private *i915,
> >   				  const u32 hpd[HPD_NUM_PINS])
> >   {
> >   	struct intel_encoder *encoder;
> >   	u32 enabled_irqs = 0;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder)
> > -		if (dev_priv->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED)
> > +	for_each_intel_encoder(&i915->drm, encoder)
> > +		if (i915->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED)
> >   			enabled_irqs |= hpd[encoder->hpd_pin];
> >   	return enabled_irqs;
> >   }
> > -static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void ibx_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug;
> > @@ -3724,29 +3724,29 @@ static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv)
> >   	 * When CPU and PCH are on the same package, port A
> >   	 * HPD must be enabled in both north and south.
> >   	 */
> > -	if (HAS_PCH_LPT_LP(dev_priv))
> > +	if (HAS_PCH_LPT_LP(i915))
> >   		hotplug |= PORTA_HOTPLUG_ENABLE;
> >   	I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
> >   }
> > -static void ibx_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void ibx_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915)) {
> >   		hotplug_irqs = SDE_HOTPLUG_MASK;
> > -		enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ibx);
> > +		enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ibx);
> >   	} else {
> >   		hotplug_irqs = SDE_HOTPLUG_MASK_CPT;
> > -		enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_cpt);
> > +		enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_cpt);
> >   	}
> > -	ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> > +	ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
> > -	ibx_hpd_detection_setup(dev_priv);
> > +	ibx_hpd_detection_setup(i915);
> >   }
> > -static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void icp_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug;
> > @@ -3763,19 +3763,19 @@ static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(SHOTPLUG_CTL_TC, hotplug);
> >   }
> > -static void icp_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void icp_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> >   	hotplug_irqs = SDE_DDI_MASK_ICP | SDE_TC_MASK_ICP;
> > -	enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_icp);
> > +	enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_icp);
> > -	ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> > +	ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
> > -	icp_hpd_detection_setup(dev_priv);
> > +	icp_hpd_detection_setup(i915);
> >   }
> > -static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void gen11_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug;
> > @@ -3794,12 +3794,12 @@ static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN11_TBT_HOTPLUG_CTL, hotplug);
> >   }
> > -static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void gen11_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> >   	u32 val;
> > -	enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_gen11);
> > +	enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_gen11);
> >   	hotplug_irqs = GEN11_DE_TC_HOTPLUG_MASK | GEN11_DE_TBT_HOTPLUG_MASK;
> >   	val = I915_READ(GEN11_DE_HPD_IMR);
> > @@ -3807,18 +3807,18 @@ static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN11_DE_HPD_IMR, val);
> >   	POSTING_READ(GEN11_DE_HPD_IMR);
> > -	gen11_hpd_detection_setup(dev_priv);
> > +	gen11_hpd_detection_setup(i915);
> > -	if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> > -		icp_hpd_irq_setup(dev_priv);
> > +	if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> > +		icp_hpd_irq_setup(i915);
> >   }
> > -static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void spt_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 val, hotplug;
> >   	/* Display WA #1179 WaHardHangonHotPlug: cnp */
> > -	if (HAS_PCH_CNP(dev_priv)) {
> > +	if (HAS_PCH_CNP(i915)) {
> >   		val = I915_READ(SOUTH_CHICKEN1);
> >   		val &= ~CHASSIS_CLK_REQ_DURATION_MASK;
> >   		val |= CHASSIS_CLK_REQ_DURATION(0xf);
> > @@ -3838,19 +3838,19 @@ static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(PCH_PORT_HOTPLUG2, hotplug);
> >   }
> > -static void spt_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void spt_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> >   	hotplug_irqs = SDE_HOTPLUG_MASK_SPT;
> > -	enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_spt);
> > +	enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_spt);
> > -	ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs);
> > +	ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs);
> > -	spt_hpd_detection_setup(dev_priv);
> > +	spt_hpd_detection_setup(i915);
> >   }
> > -static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void ilk_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug;
> > @@ -3866,33 +3866,33 @@ static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, hotplug);
> >   }
> > -static void ilk_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void ilk_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> > -	if (INTEL_GEN(dev_priv) >= 8) {
> > +	if (INTEL_GEN(i915) >= 8) {
> >   		hotplug_irqs = GEN8_PORT_DP_A_HOTPLUG;
> > -		enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bdw);
> > +		enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bdw);
> > -		bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
> > -	} else if (INTEL_GEN(dev_priv) >= 7) {
> > +		bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs);
> > +	} else if (INTEL_GEN(i915) >= 7) {
> >   		hotplug_irqs = DE_DP_A_HOTPLUG_IVB;
> > -		enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ivb);
> > +		enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ivb);
> > -		ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
> > +		ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs);
> >   	} else {
> >   		hotplug_irqs = DE_DP_A_HOTPLUG;
> > -		enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ilk);
> > +		enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ilk);
> > -		ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs);
> > +		ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs);
> >   	}
> > -	ilk_hpd_detection_setup(dev_priv);
> > +	ilk_hpd_detection_setup(i915);
> > -	ibx_hpd_irq_setup(dev_priv);
> > +	ibx_hpd_irq_setup(i915);
> >   }
> > -static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv,
> > +static void __bxt_hpd_detection_setup(struct drm_i915_private *i915,
> >   				      u32 enabled_irqs)
> >   {
> >   	u32 hotplug;
> > @@ -3911,106 +3911,106 @@ static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv,
> >   	 * for HPD detection logic, update it based on VBT fields.
> >   	 */
> >   	if ((enabled_irqs & BXT_DE_PORT_HP_DDIA) &&
> > -	    intel_bios_is_port_hpd_inverted(dev_priv, PORT_A))
> > +	    intel_bios_is_port_hpd_inverted(i915, PORT_A))
> >   		hotplug |= BXT_DDIA_HPD_INVERT;
> >   	if ((enabled_irqs & BXT_DE_PORT_HP_DDIB) &&
> > -	    intel_bios_is_port_hpd_inverted(dev_priv, PORT_B))
> > +	    intel_bios_is_port_hpd_inverted(i915, PORT_B))
> >   		hotplug |= BXT_DDIB_HPD_INVERT;
> >   	if ((enabled_irqs & BXT_DE_PORT_HP_DDIC) &&
> > -	    intel_bios_is_port_hpd_inverted(dev_priv, PORT_C))
> > +	    intel_bios_is_port_hpd_inverted(i915, PORT_C))
> >   		hotplug |= BXT_DDIC_HPD_INVERT;
> >   	I915_WRITE(PCH_PORT_HOTPLUG, hotplug);
> >   }
> > -static void bxt_hpd_detection_setup(struct drm_i915_private *dev_priv)
> > +static void bxt_hpd_detection_setup(struct drm_i915_private *i915)
> >   {
> > -	__bxt_hpd_detection_setup(dev_priv, BXT_DE_PORT_HOTPLUG_MASK);
> > +	__bxt_hpd_detection_setup(i915, BXT_DE_PORT_HOTPLUG_MASK);
> >   }
> > -static void bxt_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void bxt_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_irqs, enabled_irqs;
> > -	enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bxt);
> > +	enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bxt);
> >   	hotplug_irqs = BXT_DE_PORT_HOTPLUG_MASK;
> > -	bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs);
> > +	bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs);
> > -	__bxt_hpd_detection_setup(dev_priv, enabled_irqs);
> > +	__bxt_hpd_detection_setup(i915, enabled_irqs);
> >   }
> >   static void ibx_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 mask;
> > -	if (HAS_PCH_NOP(dev_priv))
> > +	if (HAS_PCH_NOP(i915))
> >   		return;
> > -	if (HAS_PCH_IBX(dev_priv))
> > +	if (HAS_PCH_IBX(i915))
> >   		mask = SDE_GMBUS | SDE_AUX_MASK | SDE_POISON;
> > -	else if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv))
> > +	else if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915))
> >   		mask = SDE_GMBUS_CPT | SDE_AUX_MASK_CPT;
> >   	else
> >   		mask = SDE_GMBUS_CPT;
> > -	gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR);
> > +	gen3_assert_iir_is_zero(&i915->uncore, SDEIIR);
> >   	I915_WRITE(SDEIMR, ~mask);
> > -	if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) ||
> > -	    HAS_PCH_LPT(dev_priv))
> > -		ibx_hpd_detection_setup(dev_priv);
> > +	if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) ||
> > +	    HAS_PCH_LPT(i915))
> > +		ibx_hpd_detection_setup(i915);
> >   	else
> > -		spt_hpd_detection_setup(dev_priv);
> > +		spt_hpd_detection_setup(i915);
> >   }
> >   static void gen5_gt_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 pm_irqs, gt_irqs;
> >   	pm_irqs = gt_irqs = 0;
> > -	dev_priv->gt_irq_mask = ~0;
> > -	if (HAS_L3_DPF(dev_priv)) {
> > +	i915->gt_irq_mask = ~0;
> > +	if (HAS_L3_DPF(i915)) {
> >   		/* L3 parity interrupt is always unmasked. */
> > -		dev_priv->gt_irq_mask = ~GT_PARITY_ERROR(dev_priv);
> > -		gt_irqs |= GT_PARITY_ERROR(dev_priv);
> > +		i915->gt_irq_mask = ~GT_PARITY_ERROR(i915);
> > +		gt_irqs |= GT_PARITY_ERROR(i915);
> >   	}
> >   	gt_irqs |= GT_RENDER_USER_INTERRUPT;
> > -	if (IS_GEN(dev_priv, 5)) {
> > +	if (IS_GEN(i915, 5)) {
> >   		gt_irqs |= ILK_BSD_USER_INTERRUPT;
> >   	} else {
> >   		gt_irqs |= GT_BLT_USER_INTERRUPT | GT_BSD_USER_INTERRUPT;
> >   	}
> > -	GEN3_IRQ_INIT(uncore, GT, dev_priv->gt_irq_mask, gt_irqs);
> > +	GEN3_IRQ_INIT(uncore, GT, i915->gt_irq_mask, gt_irqs);
> > -	if (INTEL_GEN(dev_priv) >= 6) {
> > +	if (INTEL_GEN(i915) >= 6) {
> >   		/*
> >   		 * RPS interrupts will get enabled/disabled on demand when RPS
> >   		 * itself is enabled/disabled.
> >   		 */
> > -		if (HAS_ENGINE(dev_priv, VECS0)) {
> > +		if (HAS_ENGINE(i915, VECS0)) {
> >   			pm_irqs |= PM_VEBOX_USER_INTERRUPT;
> > -			dev_priv->pm_ier |= PM_VEBOX_USER_INTERRUPT;
> > +			i915->pm_ier |= PM_VEBOX_USER_INTERRUPT;
> >   		}
> > -		dev_priv->pm_imr = 0xffffffff;
> > -		GEN3_IRQ_INIT(uncore, GEN6_PM, dev_priv->pm_imr, pm_irqs);
> > +		i915->pm_imr = 0xffffffff;
> > +		GEN3_IRQ_INIT(uncore, GEN6_PM, i915->pm_imr, pm_irqs);
> >   	}
> >   }
> >   static int ironlake_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 display_mask, extra_mask;
> > -	if (INTEL_GEN(dev_priv) >= 7) {
> > +	if (INTEL_GEN(i915) >= 7) {
> >   		display_mask = (DE_MASTER_IRQ_CONTROL | DE_GSE_IVB |
> >   				DE_PCH_EVENT_IVB | DE_AUX_CHANNEL_A_IVB);
> >   		extra_mask = (DE_PIPEC_VBLANK_IVB | DE_PIPEB_VBLANK_IVB |
> > @@ -4025,78 +4025,78 @@ static int ironlake_irq_postinstall(struct drm_device *dev)
> >   			      DE_DP_A_HOTPLUG);
> >   	}
> > -	if (IS_HASWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915)) {
> >   		gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR);
> > -		intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> > +		intel_psr_irq_control(i915, i915->psr.debug);
> >   		display_mask |= DE_EDP_PSR_INT_HSW;
> >   	}
> > -	dev_priv->irq_mask = ~display_mask;
> > +	i915->irq_mask = ~display_mask;
> >   	ibx_irq_pre_postinstall(dev);
> > -	GEN3_IRQ_INIT(uncore, DE, dev_priv->irq_mask,
> > +	GEN3_IRQ_INIT(uncore, DE, i915->irq_mask,
> >   		      display_mask | extra_mask);
> >   	gen5_gt_irq_postinstall(dev);
> > -	ilk_hpd_detection_setup(dev_priv);
> > +	ilk_hpd_detection_setup(i915);
> >   	ibx_irq_postinstall(dev);
> > -	if (IS_IRONLAKE_M(dev_priv)) {
> > +	if (IS_IRONLAKE_M(i915)) {
> >   		/* Enable PCU event interrupts
> >   		 *
> >   		 * spinlocking not required here for correctness since interrupt
> >   		 * setup is guaranteed to run in single-threaded context. But we
> >   		 * need it to make the assert_spin_locked happy. */
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		ilk_enable_display_irq(dev_priv, DE_PCU_EVENT);
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +		spin_lock_irq(&i915->irq_lock);
> > +		ilk_enable_display_irq(i915, DE_PCU_EVENT);
> > +		spin_unlock_irq(&i915->irq_lock);
> >   	}
> >   	return 0;
> >   }
> > -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv)
> > +void valleyview_enable_display_irqs(struct drm_i915_private *i915)
> >   {
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	if (dev_priv->display_irqs_enabled)
> > +	if (i915->display_irqs_enabled)
> >   		return;
> > -	dev_priv->display_irqs_enabled = true;
> > +	i915->display_irqs_enabled = true;
> > -	if (intel_irqs_enabled(dev_priv)) {
> > -		vlv_display_irq_reset(dev_priv);
> > -		vlv_display_irq_postinstall(dev_priv);
> > +	if (intel_irqs_enabled(i915)) {
> > +		vlv_display_irq_reset(i915);
> > +		vlv_display_irq_postinstall(i915);
> >   	}
> >   }
> > -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv)
> > +void valleyview_disable_display_irqs(struct drm_i915_private *i915)
> >   {
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	if (!dev_priv->display_irqs_enabled)
> > +	if (!i915->display_irqs_enabled)
> >   		return;
> > -	dev_priv->display_irqs_enabled = false;
> > +	i915->display_irqs_enabled = false;
> > -	if (intel_irqs_enabled(dev_priv))
> > -		vlv_display_irq_reset(dev_priv);
> > +	if (intel_irqs_enabled(i915))
> > +		vlv_display_irq_reset(i915);
> >   }
> >   static int valleyview_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	gen5_gt_irq_postinstall(dev);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->display_irqs_enabled)
> > -		vlv_display_irq_postinstall(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->display_irqs_enabled)
> > +		vlv_display_irq_postinstall(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE);
> >   	POSTING_READ(VLV_MASTER_IER);
> > @@ -4104,9 +4104,9 @@ static int valleyview_irq_postinstall(struct drm_device *dev)
> >   	return 0;
> >   }
> > -static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> > +static void gen8_gt_irq_postinstall(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	/* These are interrupts we'll toggle with the ring mask register */
> >   	u32 gt_interrupts[] = {
> > @@ -4126,21 +4126,21 @@ static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> >   		 GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VECS_IRQ_SHIFT)
> >   	};
> > -	dev_priv->pm_ier = 0x0;
> > -	dev_priv->pm_imr = ~dev_priv->pm_ier;
> > +	i915->pm_ier = 0x0;
> > +	i915->pm_imr = ~i915->pm_ier;
> >   	GEN8_IRQ_INIT_NDX(uncore, GT, 0, ~gt_interrupts[0], gt_interrupts[0]);
> >   	GEN8_IRQ_INIT_NDX(uncore, GT, 1, ~gt_interrupts[1], gt_interrupts[1]);
> >   	/*
> >   	 * RPS interrupts will get enabled/disabled on demand when RPS itself
> >   	 * is enabled/disabled. Same wil be the case for GuC interrupts.
> >   	 */
> > -	GEN8_IRQ_INIT_NDX(uncore, GT, 2, dev_priv->pm_imr, dev_priv->pm_ier);
> > +	GEN8_IRQ_INIT_NDX(uncore, GT, 2, i915->pm_imr, i915->pm_ier);
> >   	GEN8_IRQ_INIT_NDX(uncore, GT, 3, ~gt_interrupts[3], gt_interrupts[3]);
> >   }
> > -static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv)
> > +static void gen8_de_irq_postinstall(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 de_pipe_masked = GEN8_PIPE_CDCLK_CRC_DONE;
> >   	u32 de_pipe_enables;
> > @@ -4149,84 +4149,84 @@ static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv)
> >   	u32 de_misc_masked = GEN8_DE_EDP_PSR;
> >   	enum pipe pipe;
> > -	if (INTEL_GEN(dev_priv) <= 10)
> > +	if (INTEL_GEN(i915) <= 10)
> >   		de_misc_masked |= GEN8_DE_MISC_GSE;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		de_pipe_masked |= GEN9_DE_PIPE_IRQ_FAULT_ERRORS;
> >   		de_port_masked |= GEN9_AUX_CHANNEL_B | GEN9_AUX_CHANNEL_C |
> >   				  GEN9_AUX_CHANNEL_D;
> > -		if (IS_GEN9_LP(dev_priv))
> > +		if (IS_GEN9_LP(i915))
> >   			de_port_masked |= BXT_DE_PORT_GMBUS;
> >   	} else {
> >   		de_pipe_masked |= GEN8_DE_PIPE_IRQ_FAULT_ERRORS;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		de_port_masked |= ICL_AUX_CHANNEL_E;
> > -	if (IS_CNL_WITH_PORT_F(dev_priv) || INTEL_GEN(dev_priv) >= 11)
> > +	if (IS_CNL_WITH_PORT_F(i915) || INTEL_GEN(i915) >= 11)
> >   		de_port_masked |= CNL_AUX_CHANNEL_F;
> >   	de_pipe_enables = de_pipe_masked | GEN8_PIPE_VBLANK |
> >   					   GEN8_PIPE_FIFO_UNDERRUN;
> >   	de_port_enables = de_port_masked;
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		de_port_enables |= BXT_DE_PORT_HOTPLUG_MASK;
> > -	else if (IS_BROADWELL(dev_priv))
> > +	else if (IS_BROADWELL(i915))
> >   		de_port_enables |= GEN8_PORT_DP_A_HOTPLUG;
> >   	gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR);
> > -	intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> > +	intel_psr_irq_control(i915, i915->psr.debug);
> > -	for_each_pipe(dev_priv, pipe) {
> > -		dev_priv->de_irq_mask[pipe] = ~de_pipe_masked;
> > +	for_each_pipe(i915, pipe) {
> > +		i915->de_irq_mask[pipe] = ~de_pipe_masked;
> > -		if (intel_display_power_is_enabled(dev_priv,
> > +		if (intel_display_power_is_enabled(i915,
> >   				POWER_DOMAIN_PIPE(pipe)))
> >   			GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe,
> > -					  dev_priv->de_irq_mask[pipe],
> > +					  i915->de_irq_mask[pipe],
> >   					  de_pipe_enables);
> >   	}
> >   	GEN3_IRQ_INIT(uncore, GEN8_DE_PORT_, ~de_port_masked, de_port_enables);
> >   	GEN3_IRQ_INIT(uncore, GEN8_DE_MISC_, ~de_misc_masked, de_misc_masked);
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		u32 de_hpd_masked = 0;
> >   		u32 de_hpd_enables = GEN11_DE_TC_HOTPLUG_MASK |
> >   				     GEN11_DE_TBT_HOTPLUG_MASK;
> >   		GEN3_IRQ_INIT(uncore, GEN11_DE_HPD_, ~de_hpd_masked,
> >   			      de_hpd_enables);
> > -		gen11_hpd_detection_setup(dev_priv);
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > -		bxt_hpd_detection_setup(dev_priv);
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		ilk_hpd_detection_setup(dev_priv);
> > +		gen11_hpd_detection_setup(i915);
> > +	} else if (IS_GEN9_LP(i915)) {
> > +		bxt_hpd_detection_setup(i915);
> > +	} else if (IS_BROADWELL(i915)) {
> > +		ilk_hpd_detection_setup(i915);
> >   	}
> >   }
> >   static int gen8_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		ibx_irq_pre_postinstall(dev);
> > -	gen8_gt_irq_postinstall(dev_priv);
> > -	gen8_de_irq_postinstall(dev_priv);
> > +	gen8_gt_irq_postinstall(i915);
> > +	gen8_de_irq_postinstall(i915);
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		ibx_irq_postinstall(dev);
> > -	gen8_master_intr_enable(dev_priv->uncore.regs);
> > +	gen8_master_intr_enable(i915->uncore.regs);
> >   	return 0;
> >   }
> > -static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> > +static void gen11_gt_irq_postinstall(struct drm_i915_private *i915)
> >   {
> >   	const u32 irqs = GT_RENDER_USER_INTERRUPT | GT_CONTEXT_SWITCH_INTERRUPT;
> > @@ -4247,8 +4247,8 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> >   	 * RPS interrupts will get enabled/disabled on demand when RPS itself
> >   	 * is enabled/disabled.
> >   	 */
> > -	dev_priv->pm_ier = 0x0;
> > -	dev_priv->pm_imr = ~dev_priv->pm_ier;
> > +	i915->pm_ier = 0x0;
> > +	i915->pm_imr = ~i915->pm_ier;
> >   	I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_ENABLE, 0);
> >   	I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_MASK,  ~0);
> > @@ -4259,36 +4259,36 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv)
> >   static void icp_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 mask = SDE_GMBUS_ICP;
> >   	WARN_ON(I915_READ(SDEIER) != 0);
> >   	I915_WRITE(SDEIER, 0xffffffff);
> >   	POSTING_READ(SDEIER);
> > -	gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR);
> > +	gen3_assert_iir_is_zero(&i915->uncore, SDEIIR);
> >   	I915_WRITE(SDEIMR, ~mask);
> > -	icp_hpd_detection_setup(dev_priv);
> > +	icp_hpd_detection_setup(i915);
> >   }
> >   static int gen11_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = dev->dev_private;
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = dev->dev_private;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 gu_misc_masked = GEN11_GU_MISC_GSE;
> > -	if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> > +	if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> >   		icp_irq_postinstall(dev);
> > -	gen11_gt_irq_postinstall(dev_priv);
> > -	gen8_de_irq_postinstall(dev_priv);
> > +	gen11_gt_irq_postinstall(i915);
> > +	gen8_de_irq_postinstall(i915);
> >   	GEN3_IRQ_INIT(uncore, GEN11_GU_MISC_, ~gu_misc_masked, gu_misc_masked);
> >   	I915_WRITE(GEN11_DISPLAY_INT_CTL, GEN11_DISPLAY_IRQ_ENABLE);
> > -	gen11_master_intr_enable(dev_priv->uncore.regs);
> > +	gen11_master_intr_enable(i915->uncore.regs);
> >   	POSTING_READ(GEN11_GFX_MSTR_IRQ);
> >   	return 0;
> > @@ -4296,14 +4296,14 @@ static int gen11_irq_postinstall(struct drm_device *dev)
> >   static int cherryview_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	gen8_gt_irq_postinstall(dev_priv);
> > +	gen8_gt_irq_postinstall(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->display_irqs_enabled)
> > -		vlv_display_irq_postinstall(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->display_irqs_enabled)
> > +		vlv_display_irq_postinstall(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL);
> >   	POSTING_READ(GEN8_MASTER_IRQ);
> > @@ -4313,18 +4313,18 @@ static int cherryview_irq_postinstall(struct drm_device *dev)
> >   static void i8xx_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	i9xx_pipestat_irq_reset(dev_priv);
> > +	i9xx_pipestat_irq_reset(i915);
> >   	GEN2_IRQ_RESET(uncore);
> >   }
> >   static int i8xx_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u16 enable_mask;
> >   	intel_uncore_write16(uncore,
> > @@ -4333,7 +4333,7 @@ static int i8xx_irq_postinstall(struct drm_device *dev)
> >   			       I915_ERROR_MEMORY_REFRESH));
> >   	/* Unmask the interrupts that we always want on. */
> > -	dev_priv->irq_mask =
> > +	i915->irq_mask =
> >   		~(I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> >   		  I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
> >   		  I915_MASTER_ERROR_INTERRUPT);
> > @@ -4344,14 +4344,14 @@ static int i8xx_irq_postinstall(struct drm_device *dev)
> >   		I915_MASTER_ERROR_INTERRUPT |
> >   		I915_USER_INTERRUPT;
> > -	GEN2_IRQ_INIT(uncore, dev_priv->irq_mask, enable_mask);
> > +	GEN2_IRQ_INIT(uncore, i915->irq_mask, enable_mask);
> >   	/* Interrupt setup is already guaranteed to be single-threaded, this is
> >   	 * just to make the assert_spin_locked check happy. */
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	return 0;
> >   }
> > @@ -4386,7 +4386,7 @@ static void i8xx_error_irq_ack(struct drm_i915_private *i915,
> >   	intel_uncore_write16(uncore, EMR, emr | *eir_stuck);
> >   }
> > -static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i8xx_error_irq_handler(struct drm_i915_private *i915,
> >   				   u16 eir, u16 eir_stuck)
> >   {
> >   	DRM_DEBUG("Master Error: EIR 0x%04x\n", eir);
> > @@ -4395,7 +4395,7 @@ static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv,
> >   		DRM_DEBUG_DRIVER("EIR stuck: 0x%04x, masked\n", eir_stuck);
> >   }
> > -static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv,
> > +static void i9xx_error_irq_ack(struct drm_i915_private *i915,
> >   			       u32 *eir, u32 *eir_stuck)
> >   {
> >   	u32 emr;
> > @@ -4423,7 +4423,7 @@ static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(EMR, emr | *eir_stuck);
> >   }
> > -static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv,
> > +static void i9xx_error_irq_handler(struct drm_i915_private *i915,
> >   				   u32 eir, u32 eir_stuck)
> >   {
> >   	DRM_DEBUG("Master Error, EIR 0x%08x\n", eir);
> > @@ -4435,21 +4435,21 @@ static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv,
> >   static irqreturn_t i8xx_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	do {
> >   		u32 pipe_stats[I915_MAX_PIPES] = {};
> >   		u16 eir = 0, eir_stuck = 0;
> >   		u16 iir;
> > -		iir = intel_uncore_read16(&dev_priv->uncore, GEN2_IIR);
> > +		iir = intel_uncore_read16(&i915->uncore, GEN2_IIR);
> >   		if (iir == 0)
> >   			break;
> > @@ -4457,53 +4457,53 @@ static irqreturn_t i8xx_irq_handler(int irq, void *arg)
> >   		/* Call regardless, as some status bits might not be
> >   		 * signalled in iir */
> > -		i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> > +		i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i8xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> > +			i8xx_error_irq_ack(i915, &eir, &eir_stuck);
> > -		intel_uncore_write16(&dev_priv->uncore, GEN2_IIR, iir);
> > +		intel_uncore_write16(&i915->uncore, GEN2_IIR, iir);
> >   		if (iir & I915_USER_INTERRUPT)
> > -			intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> > +			intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i8xx_error_irq_handler(dev_priv, eir, eir_stuck);
> > +			i8xx_error_irq_handler(i915, eir, eir_stuck);
> > -		i8xx_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> > +		i8xx_pipestat_irq_handler(i915, iir, pipe_stats);
> >   	} while (0);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> >   static void i915_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	if (I915_HAS_HOTPLUG(dev_priv)) {
> > -		i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
> > +	if (I915_HAS_HOTPLUG(i915)) {
> > +		i915_hotplug_interrupt_update(i915, 0xffffffff, 0);
> >   		I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
> >   	}
> > -	i9xx_pipestat_irq_reset(dev_priv);
> > +	i9xx_pipestat_irq_reset(i915);
> >   	GEN3_IRQ_RESET(uncore, GEN2_);
> >   }
> >   static int i915_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 enable_mask;
> >   	I915_WRITE(EMR, ~(I915_ERROR_PAGE_TABLE |
> >   			  I915_ERROR_MEMORY_REFRESH));
> >   	/* Unmask the interrupts that we always want on. */
> > -	dev_priv->irq_mask =
> > +	i915->irq_mask =
> >   		~(I915_ASLE_INTERRUPT |
> >   		  I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> >   		  I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |
> > @@ -4516,23 +4516,23 @@ static int i915_irq_postinstall(struct drm_device *dev)
> >   		I915_MASTER_ERROR_INTERRUPT |
> >   		I915_USER_INTERRUPT;
> > -	if (I915_HAS_HOTPLUG(dev_priv)) {
> > +	if (I915_HAS_HOTPLUG(i915)) {
> >   		/* Enable in IER... */
> >   		enable_mask |= I915_DISPLAY_PORT_INTERRUPT;
> >   		/* and unmask in IMR */
> > -		dev_priv->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT;
> > +		i915->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT;
> >   	}
> > -	GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask);
> > +	GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask);
> >   	/* Interrupt setup is already guaranteed to be single-threaded, this is
> >   	 * just to make the assert_spin_locked check happy. */
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	i915_enable_asle_pipestat(dev_priv);
> > +	i915_enable_asle_pipestat(i915);
> >   	return 0;
> >   }
> > @@ -4540,14 +4540,14 @@ static int i915_irq_postinstall(struct drm_device *dev)
> >   static irqreturn_t i915_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	do {
> >   		u32 pipe_stats[I915_MAX_PIPES] = {};
> > @@ -4561,53 +4561,53 @@ static irqreturn_t i915_irq_handler(int irq, void *arg)
> >   		ret = IRQ_HANDLED;
> > -		if (I915_HAS_HOTPLUG(dev_priv) &&
> > +		if (I915_HAS_HOTPLUG(i915) &&
> >   		    iir & I915_DISPLAY_PORT_INTERRUPT)
> > -			hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> > +			hotplug_status = i9xx_hpd_irq_ack(i915);
> >   		/* Call regardless, as some status bits might not be
> >   		 * signalled in iir */
> > -		i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> > +		i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> > +			i9xx_error_irq_ack(i915, &eir, &eir_stuck);
> >   		I915_WRITE(GEN2_IIR, iir);
> >   		if (iir & I915_USER_INTERRUPT)
> > -			intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> > +			intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i9xx_error_irq_handler(dev_priv, eir, eir_stuck);
> > +			i9xx_error_irq_handler(i915, eir, eir_stuck);
> >   		if (hotplug_status)
> > -			i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> > +			i9xx_hpd_irq_handler(i915, hotplug_status);
> > -		i915_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> > +		i915_pipestat_irq_handler(i915, iir, pipe_stats);
> >   	} while (0);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> >   static void i965_irq_reset(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0);
> > +	i915_hotplug_interrupt_update(i915, 0xffffffff, 0);
> >   	I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
> > -	i9xx_pipestat_irq_reset(dev_priv);
> > +	i9xx_pipestat_irq_reset(i915);
> >   	GEN3_IRQ_RESET(uncore, GEN2_);
> >   }
> >   static int i965_irq_postinstall(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 enable_mask;
> >   	u32 error_mask;
> > @@ -4615,7 +4615,7 @@ static int i965_irq_postinstall(struct drm_device *dev)
> >   	 * Enable some error detection, note the instruction error mask
> >   	 * bit is reserved, so we leave it masked.
> >   	 */
> > -	if (IS_G4X(dev_priv)) {
> > +	if (IS_G4X(i915)) {
> >   		error_mask = ~(GM45_ERROR_PAGE_TABLE |
> >   			       GM45_ERROR_MEM_PRIV |
> >   			       GM45_ERROR_CP_PRIV |
> > @@ -4627,7 +4627,7 @@ static int i965_irq_postinstall(struct drm_device *dev)
> >   	I915_WRITE(EMR, error_mask);
> >   	/* Unmask the interrupts that we always want on. */
> > -	dev_priv->irq_mask =
> > +	i915->irq_mask =
> >   		~(I915_ASLE_INTERRUPT |
> >   		  I915_DISPLAY_PORT_INTERRUPT |
> >   		  I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |
> > @@ -4642,43 +4642,43 @@ static int i965_irq_postinstall(struct drm_device *dev)
> >   		I915_MASTER_ERROR_INTERRUPT |
> >   		I915_USER_INTERRUPT;
> > -	if (IS_G4X(dev_priv))
> > +	if (IS_G4X(i915))
> >   		enable_mask |= I915_BSD_USER_INTERRUPT;
> > -	GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask);
> > +	GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask);
> >   	/* Interrupt setup is already guaranteed to be single-threaded, this is
> >   	 * just to make the assert_spin_locked check happy. */
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> > -	i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS);
> > +	i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS);
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	i915_enable_asle_pipestat(dev_priv);
> > +	i915_enable_asle_pipestat(i915);
> >   	return 0;
> >   }
> > -static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv)
> > +static void i915_hpd_irq_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 hotplug_en;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	/* Note HDMI and DP share hotplug bits */
> >   	/* enable bits are the same for all generations */
> > -	hotplug_en = intel_hpd_enabled_irqs(dev_priv, hpd_mask_i915);
> > +	hotplug_en = intel_hpd_enabled_irqs(i915, hpd_mask_i915);
> >   	/* Programming the CRT detection parameters tends
> >   	   to generate a spurious hotplug event about three
> >   	   seconds later.  So just do it once.
> >   	*/
> > -	if (IS_G4X(dev_priv))
> > +	if (IS_G4X(i915))
> >   		hotplug_en |= CRT_HOTPLUG_ACTIVATION_PERIOD_64;
> >   	hotplug_en |= CRT_HOTPLUG_VOLTAGE_COMPARE_50;
> >   	/* Ignore TV since it's buggy */
> > -	i915_hotplug_interrupt_update_locked(dev_priv,
> > +	i915_hotplug_interrupt_update_locked(i915,
> >   					     HOTPLUG_INT_EN_MASK |
> >   					     CRT_HOTPLUG_VOLTAGE_COMPARE_MASK |
> >   					     CRT_HOTPLUG_ACTIVATION_PERIOD_64,
> > @@ -4688,14 +4688,14 @@ static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv)
> >   static irqreturn_t i965_irq_handler(int irq, void *arg)
> >   {
> >   	struct drm_device *dev = arg;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	irqreturn_t ret = IRQ_NONE;
> > -	if (!intel_irqs_enabled(dev_priv))
> > +	if (!intel_irqs_enabled(i915))
> >   		return IRQ_NONE;
> >   	/* IRQs are synced during runtime_suspend, we don't require a wakeref */
> > -	disable_rpm_wakeref_asserts(dev_priv);
> > +	disable_rpm_wakeref_asserts(i915);
> >   	do {
> >   		u32 pipe_stats[I915_MAX_PIPES] = {};
> > @@ -4710,76 +4710,76 @@ static irqreturn_t i965_irq_handler(int irq, void *arg)
> >   		ret = IRQ_HANDLED;
> >   		if (iir & I915_DISPLAY_PORT_INTERRUPT)
> > -			hotplug_status = i9xx_hpd_irq_ack(dev_priv);
> > +			hotplug_status = i9xx_hpd_irq_ack(i915);
> >   		/* Call regardless, as some status bits might not be
> >   		 * signalled in iir */
> > -		i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats);
> > +		i9xx_pipestat_irq_ack(i915, iir, pipe_stats);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck);
> > +			i9xx_error_irq_ack(i915, &eir, &eir_stuck);
> >   		I915_WRITE(GEN2_IIR, iir);
> >   		if (iir & I915_USER_INTERRUPT)
> > -			intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]);
> > +			intel_engine_breadcrumbs_irq(i915->engine[RCS0]);
> >   		if (iir & I915_BSD_USER_INTERRUPT)
> > -			intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]);
> > +			intel_engine_breadcrumbs_irq(i915->engine[VCS0]);
> >   		if (iir & I915_MASTER_ERROR_INTERRUPT)
> > -			i9xx_error_irq_handler(dev_priv, eir, eir_stuck);
> > +			i9xx_error_irq_handler(i915, eir, eir_stuck);
> >   		if (hotplug_status)
> > -			i9xx_hpd_irq_handler(dev_priv, hotplug_status);
> > +			i9xx_hpd_irq_handler(i915, hotplug_status);
> > -		i965_pipestat_irq_handler(dev_priv, iir, pipe_stats);
> > +		i965_pipestat_irq_handler(i915, iir, pipe_stats);
> >   	} while (0);
> > -	enable_rpm_wakeref_asserts(dev_priv);
> > +	enable_rpm_wakeref_asserts(i915);
> >   	return ret;
> >   }
> >   /**
> >    * intel_irq_init - initializes irq support
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function initializes all the irq support including work items, timers
> >    * and all the vtables. It does not setup the interrupt itself though.
> >    */
> > -void intel_irq_init(struct drm_i915_private *dev_priv)
> > +void intel_irq_init(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_device *dev = &i915->drm;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	int i;
> > -	if (IS_I945GM(dev_priv))
> > -		i945gm_vblank_work_init(dev_priv);
> > +	if (IS_I945GM(i915))
> > +		i945gm_vblank_work_init(i915);
> > -	intel_hpd_init_work(dev_priv);
> > +	intel_hpd_init_work(i915);
> >   	INIT_WORK(&rps->work, gen6_pm_rps_work);
> > -	INIT_WORK(&dev_priv->l3_parity.error_work, ivybridge_parity_work);
> > +	INIT_WORK(&i915->l3_parity.error_work, ivybridge_parity_work);
> >   	for (i = 0; i < MAX_L3_SLICES; ++i)
> > -		dev_priv->l3_parity.remap_info[i] = NULL;
> > +		i915->l3_parity.remap_info[i] = NULL;
> > -	if (HAS_GUC_SCHED(dev_priv) && INTEL_GEN(dev_priv) < 11)
> > -		dev_priv->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT;
> > +	if (HAS_GUC_SCHED(i915) && INTEL_GEN(i915) < 11)
> > +		i915->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT;
> >   	/* Let's track the enabled rps events */
> > -	if (IS_VALLEYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915))
> >   		/* WaGsvRC0ResidencyMethod:vlv */
> > -		dev_priv->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED;
> > +		i915->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED;
> >   	else
> > -		dev_priv->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD |
> > +		i915->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD |
> >   					   GEN6_PM_RP_DOWN_THRESHOLD |
> >   					   GEN6_PM_RP_DOWN_TIMEOUT);
> >   	/* We share the register with other engine */
> > -	if (INTEL_GEN(dev_priv) > 9)
> > -		GEM_WARN_ON(dev_priv->pm_rps_events & 0xffff0000);
> > +	if (INTEL_GEN(i915) > 9)
> > +		GEM_WARN_ON(i915->pm_rps_events & 0xffff0000);
> >   	rps->pm_intrmsk_mbz = 0;
> > @@ -4789,15 +4789,15 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> >   	 *
> >   	 * TODO: verify if this can be reproduced on VLV,CHV.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) <= 7)
> > +	if (INTEL_GEN(i915) <= 7)
> >   		rps->pm_intrmsk_mbz |= GEN6_PM_RP_UP_EI_EXPIRED;
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		rps->pm_intrmsk_mbz |= GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC;
> > -	if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> >   		dev->driver->get_vblank_counter = g4x_get_vblank_counter;
> > -	else if (INTEL_GEN(dev_priv) >= 3)
> > +	else if (INTEL_GEN(i915) >= 3)
> >   		dev->driver->get_vblank_counter = i915_get_vblank_counter;
> >   	dev->vblank_disable_immediate = true;
> > @@ -4808,83 +4808,83 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> >   	 * outside of the power domain. We defer setting up the display irqs
> >   	 * in this case to the runtime pm.
> >   	 */
> > -	dev_priv->display_irqs_enabled = true;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->display_irqs_enabled = false;
> > +	i915->display_irqs_enabled = true;
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->display_irqs_enabled = false;
> > -	dev_priv->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD;
> > +	i915->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD;
> >   	/* If we have MST support, we want to avoid doing short HPD IRQ storm
> >   	 * detection, as short HPD storms will occur as a natural part of
> >   	 * sideband messaging with MST.
> >   	 * On older platforms however, IRQ storms can occur with both long and
> >   	 * short pulses, as seen on some G4x systems.
> >   	 */
> > -	dev_priv->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(dev_priv);
> > +	i915->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(i915);
> >   	dev->driver->get_vblank_timestamp = drm_calc_vbltimestamp_from_scanoutpos;
> >   	dev->driver->get_scanout_position = i915_get_crtc_scanoutpos;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		dev->driver->irq_handler = cherryview_irq_handler;
> >   		dev->driver->irq_preinstall = cherryview_irq_reset;
> >   		dev->driver->irq_postinstall = cherryview_irq_postinstall;
> >   		dev->driver->irq_uninstall = cherryview_irq_reset;
> >   		dev->driver->enable_vblank = i965_enable_vblank;
> >   		dev->driver->disable_vblank = i965_disable_vblank;
> > -		dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +		i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		dev->driver->irq_handler = valleyview_irq_handler;
> >   		dev->driver->irq_preinstall = valleyview_irq_reset;
> >   		dev->driver->irq_postinstall = valleyview_irq_postinstall;
> >   		dev->driver->irq_uninstall = valleyview_irq_reset;
> >   		dev->driver->enable_vblank = i965_enable_vblank;
> >   		dev->driver->disable_vblank = i965_disable_vblank;
> > -		dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> > -	} else if (INTEL_GEN(dev_priv) >= 11) {
> > +		i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> > +	} else if (INTEL_GEN(i915) >= 11) {
> >   		dev->driver->irq_handler = gen11_irq_handler;
> >   		dev->driver->irq_preinstall = gen11_irq_reset;
> >   		dev->driver->irq_postinstall = gen11_irq_postinstall;
> >   		dev->driver->irq_uninstall = gen11_irq_reset;
> >   		dev->driver->enable_vblank = gen8_enable_vblank;
> >   		dev->driver->disable_vblank = gen8_disable_vblank;
> > -		dev_priv->display.hpd_irq_setup = gen11_hpd_irq_setup;
> > -	} else if (INTEL_GEN(dev_priv) >= 8) {
> > +		i915->display.hpd_irq_setup = gen11_hpd_irq_setup;
> > +	} else if (INTEL_GEN(i915) >= 8) {
> >   		dev->driver->irq_handler = gen8_irq_handler;
> >   		dev->driver->irq_preinstall = gen8_irq_reset;
> >   		dev->driver->irq_postinstall = gen8_irq_postinstall;
> >   		dev->driver->irq_uninstall = gen8_irq_reset;
> >   		dev->driver->enable_vblank = gen8_enable_vblank;
> >   		dev->driver->disable_vblank = gen8_disable_vblank;
> > -		if (IS_GEN9_LP(dev_priv))
> > -			dev_priv->display.hpd_irq_setup = bxt_hpd_irq_setup;
> > -		else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT)
> > -			dev_priv->display.hpd_irq_setup = spt_hpd_irq_setup;
> > +		if (IS_GEN9_LP(i915))
> > +			i915->display.hpd_irq_setup = bxt_hpd_irq_setup;
> > +		else if (INTEL_PCH_TYPE(i915) >= PCH_SPT)
> > +			i915->display.hpd_irq_setup = spt_hpd_irq_setup;
> >   		else
> > -			dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +			i915->display.hpd_irq_setup = ilk_hpd_irq_setup;
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		dev->driver->irq_handler = ironlake_irq_handler;
> >   		dev->driver->irq_preinstall = ironlake_irq_reset;
> >   		dev->driver->irq_postinstall = ironlake_irq_postinstall;
> >   		dev->driver->irq_uninstall = ironlake_irq_reset;
> >   		dev->driver->enable_vblank = ironlake_enable_vblank;
> >   		dev->driver->disable_vblank = ironlake_disable_vblank;
> > -		dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup;
> > +		i915->display.hpd_irq_setup = ilk_hpd_irq_setup;
> >   	} else {
> > -		if (IS_GEN(dev_priv, 2)) {
> > +		if (IS_GEN(i915, 2)) {
> >   			dev->driver->irq_preinstall = i8xx_irq_reset;
> >   			dev->driver->irq_postinstall = i8xx_irq_postinstall;
> >   			dev->driver->irq_handler = i8xx_irq_handler;
> >   			dev->driver->irq_uninstall = i8xx_irq_reset;
> >   			dev->driver->enable_vblank = i8xx_enable_vblank;
> >   			dev->driver->disable_vblank = i8xx_disable_vblank;
> > -		} else if (IS_I945GM(dev_priv)) {
> > +		} else if (IS_I945GM(i915)) {
> >   			dev->driver->irq_preinstall = i915_irq_reset;
> >   			dev->driver->irq_postinstall = i915_irq_postinstall;
> >   			dev->driver->irq_uninstall = i915_irq_reset;
> >   			dev->driver->irq_handler = i915_irq_handler;
> >   			dev->driver->enable_vblank = i945gm_enable_vblank;
> >   			dev->driver->disable_vblank = i945gm_disable_vblank;
> > -		} else if (IS_GEN(dev_priv, 3)) {
> > +		} else if (IS_GEN(i915, 3)) {
> >   			dev->driver->irq_preinstall = i915_irq_reset;
> >   			dev->driver->irq_postinstall = i915_irq_postinstall;
> >   			dev->driver->irq_uninstall = i915_irq_reset;
> > @@ -4899,8 +4899,8 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
> >   			dev->driver->enable_vblank = i965_enable_vblank;
> >   			dev->driver->disable_vblank = i965_disable_vblank;
> >   		}
> > -		if (I915_HAS_HOTPLUG(dev_priv))
> > -			dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup;
> > +		if (I915_HAS_HOTPLUG(i915))
> > +			i915->display.hpd_irq_setup = i915_hpd_irq_setup;
> >   	}
> >   }
> > @@ -4923,7 +4923,7 @@ void intel_irq_fini(struct drm_i915_private *i915)
> >   /**
> >    * intel_irq_install - enables the hardware interrupt
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function enables the hardware interrupt handling, but leaves the hotplug
> >    * handling still disabled. It is called after intel_irq_init().
> > @@ -4932,56 +4932,56 @@ void intel_irq_fini(struct drm_i915_private *i915)
> >    * but don't want to deal with the hassle of concurrent probe and hotplug
> >    * workers. Hence the split into this two-stage approach.
> >    */
> > -int intel_irq_install(struct drm_i915_private *dev_priv)
> > +int intel_irq_install(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * We enable some interrupt sources in our postinstall hooks, so mark
> >   	 * interrupts as enabled _before_ actually enabling them to avoid
> >   	 * special cases in our ordering checks.
> >   	 */
> > -	dev_priv->runtime_pm.irqs_enabled = true;
> > +	i915->runtime_pm.irqs_enabled = true;
> > -	return drm_irq_install(&dev_priv->drm, dev_priv->drm.pdev->irq);
> > +	return drm_irq_install(&i915->drm, i915->drm.pdev->irq);
> >   }
> >   /**
> >    * intel_irq_uninstall - finilizes all irq handling
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This stops interrupt and hotplug handling and unregisters and frees all
> >    * resources acquired in the init functions.
> >    */
> > -void intel_irq_uninstall(struct drm_i915_private *dev_priv)
> > +void intel_irq_uninstall(struct drm_i915_private *i915)
> >   {
> > -	drm_irq_uninstall(&dev_priv->drm);
> > -	intel_hpd_cancel_work(dev_priv);
> > -	dev_priv->runtime_pm.irqs_enabled = false;
> > +	drm_irq_uninstall(&i915->drm);
> > +	intel_hpd_cancel_work(i915);
> > +	i915->runtime_pm.irqs_enabled = false;
> >   }
> >   /**
> >    * intel_runtime_pm_disable_interrupts - runtime interrupt disabling
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function is used to disable interrupts at runtime, both in the runtime
> >    * pm and the system suspend/resume code.
> >    */
> > -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv)
> > +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->drm.driver->irq_uninstall(&dev_priv->drm);
> > -	dev_priv->runtime_pm.irqs_enabled = false;
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	i915->drm.driver->irq_uninstall(&i915->drm);
> > +	i915->runtime_pm.irqs_enabled = false;
> > +	synchronize_irq(i915->drm.irq);
> >   }
> >   /**
> >    * intel_runtime_pm_enable_interrupts - runtime interrupt enabling
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function is used to enable interrupts at runtime, both in the runtime
> >    * pm and the system suspend/resume code.
> >    */
> > -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv)
> > +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->runtime_pm.irqs_enabled = true;
> > -	dev_priv->drm.driver->irq_preinstall(&dev_priv->drm);
> > -	dev_priv->drm.driver->irq_postinstall(&dev_priv->drm);
> > +	i915->runtime_pm.irqs_enabled = true;
> > +	i915->drm.driver->irq_preinstall(&i915->drm);
> > +	i915->drm.driver->irq_postinstall(&i915->drm);
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_irq.h b/drivers/gpu/drm/i915/i915_irq.h
> > index cb25dd213308..604504993513 100644
> > --- a/drivers/gpu/drm/i915/i915_irq.h
> > +++ b/drivers/gpu/drm/i915/i915_irq.h
> > @@ -13,77 +13,77 @@
> >   struct drm_i915_private;
> >   struct intel_crtc;
> > -extern void intel_irq_init(struct drm_i915_private *dev_priv);
> > -extern void intel_irq_fini(struct drm_i915_private *dev_priv);
> > -int intel_irq_install(struct drm_i915_private *dev_priv);
> > -void intel_irq_uninstall(struct drm_i915_private *dev_priv);
> > +extern void intel_irq_init(struct drm_i915_private *i915);
> > +extern void intel_irq_fini(struct drm_i915_private *i915);
> > +int intel_irq_install(struct drm_i915_private *i915);
> > +void intel_irq_uninstall(struct drm_i915_private *i915);
> > -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv,
> > +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915,
> >   			      enum pipe pipe);
> >   void
> > -i915_enable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe,
> > +i915_enable_pipestat(struct drm_i915_private *i915, enum pipe pipe,
> >   		     u32 status_mask);
> >   void
> > -i915_disable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe,
> > +i915_disable_pipestat(struct drm_i915_private *i915, enum pipe pipe,
> >   		      u32 status_mask);
> > -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv);
> > -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv);
> > +void valleyview_enable_display_irqs(struct drm_i915_private *i915);
> > +void valleyview_disable_display_irqs(struct drm_i915_private *i915);
> > -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv,
> > +void i915_hotplug_interrupt_update(struct drm_i915_private *i915,
> >   				   u32 mask,
> >   				   u32 bits);
> > -void ilk_update_display_irq(struct drm_i915_private *dev_priv,
> > +void ilk_update_display_irq(struct drm_i915_private *i915,
> >   			    u32 interrupt_mask,
> >   			    u32 enabled_irq_mask);
> >   static inline void
> > -ilk_enable_display_irq(struct drm_i915_private *dev_priv, u32 bits)
> > +ilk_enable_display_irq(struct drm_i915_private *i915, u32 bits)
> >   {
> > -	ilk_update_display_irq(dev_priv, bits, bits);
> > +	ilk_update_display_irq(i915, bits, bits);
> >   }
> >   static inline void
> > -ilk_disable_display_irq(struct drm_i915_private *dev_priv, u32 bits)
> > +ilk_disable_display_irq(struct drm_i915_private *i915, u32 bits)
> >   {
> > -	ilk_update_display_irq(dev_priv, bits, 0);
> > +	ilk_update_display_irq(i915, bits, 0);
> >   }
> > -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv,
> > +void bdw_update_pipe_irq(struct drm_i915_private *i915,
> >   			 enum pipe pipe,
> >   			 u32 interrupt_mask,
> >   			 u32 enabled_irq_mask);
> > -static inline void bdw_enable_pipe_irq(struct drm_i915_private *dev_priv,
> > +static inline void bdw_enable_pipe_irq(struct drm_i915_private *i915,
> >   				       enum pipe pipe, u32 bits)
> >   {
> > -	bdw_update_pipe_irq(dev_priv, pipe, bits, bits);
> > +	bdw_update_pipe_irq(i915, pipe, bits, bits);
> >   }
> > -static inline void bdw_disable_pipe_irq(struct drm_i915_private *dev_priv,
> > +static inline void bdw_disable_pipe_irq(struct drm_i915_private *i915,
> >   					enum pipe pipe, u32 bits)
> >   {
> > -	bdw_update_pipe_irq(dev_priv, pipe, bits, 0);
> > +	bdw_update_pipe_irq(i915, pipe, bits, 0);
> >   }
> > -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv,
> > +void ibx_display_interrupt_update(struct drm_i915_private *i915,
> >   				  u32 interrupt_mask,
> >   				  u32 enabled_irq_mask);
> >   static inline void
> > -ibx_enable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits)
> > +ibx_enable_display_interrupt(struct drm_i915_private *i915, u32 bits)
> >   {
> > -	ibx_display_interrupt_update(dev_priv, bits, bits);
> > +	ibx_display_interrupt_update(i915, bits, bits);
> >   }
> >   static inline void
> > -ibx_disable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits)
> > +ibx_disable_display_interrupt(struct drm_i915_private *i915, u32 bits)
> >   {
> > -	ibx_display_interrupt_update(dev_priv, bits, 0);
> > +	ibx_display_interrupt_update(i915, bits, 0);
> >   }
> > -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask);
> > -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask);
> > -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask);
> > -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask);
> > -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv);
> > -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv);
> > -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv);
> > -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv);
> > -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv);
> > +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask);
> > +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask);
> > +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask);
> > +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask);
> > +void gen11_reset_rps_interrupts(struct drm_i915_private *i915);
> > +void gen6_reset_rps_interrupts(struct drm_i915_private *i915);
> > +void gen6_enable_rps_interrupts(struct drm_i915_private *i915);
> > +void gen6_disable_rps_interrupts(struct drm_i915_private *i915);
> > +void gen6_rps_reset_ei(struct drm_i915_private *i915);
> >   static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915,
> >   					    u32 mask)
> > @@ -91,25 +91,25 @@ static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915,
> >   	return mask & ~i915->gt_pm.rps.pm_intrmsk_mbz;
> >   }
> > -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv);
> > -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv);
> > -static inline bool intel_irqs_enabled(struct drm_i915_private *dev_priv)
> > +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915);
> > +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915);
> > +static inline bool intel_irqs_enabled(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * We only use drm_irq_uninstall() at unload and VT switch, so
> >   	 * this is the only thing we need to check.
> >   	 */
> > -	return dev_priv->runtime_pm.irqs_enabled;
> > +	return i915->runtime_pm.irqs_enabled;
> >   }
> >   int intel_get_crtc_scanline(struct intel_crtc *crtc);
> > -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
> > +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915,
> >   				     u8 pipe_mask);
> > -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv,
> > +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915,
> >   				     u8 pipe_mask);
> > -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv);
> > -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv);
> > -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv);
> > +void gen9_reset_guc_interrupts(struct drm_i915_private *i915);
> > +void gen9_enable_guc_interrupts(struct drm_i915_private *i915);
> > +void gen9_disable_guc_interrupts(struct drm_i915_private *i915);
> >   void gen11_reset_guc_interrupts(struct drm_i915_private *i915);
> >   void gen11_enable_guc_interrupts(struct drm_i915_private *i915);
> >   void gen11_disable_guc_interrupts(struct drm_i915_private *i915);
> > diff --git a/drivers/gpu/drm/i915/i915_memcpy.c b/drivers/gpu/drm/i915/i915_memcpy.c
> > index 79f8ec756362..2bd664c16ba0 100644
> > --- a/drivers/gpu/drm/i915/i915_memcpy.c
> > +++ b/drivers/gpu/drm/i915/i915_memcpy.c
> > @@ -94,7 +94,7 @@ bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len)
> >   	return false;
> >   }
> > -void i915_memcpy_init_early(struct drm_i915_private *dev_priv)
> > +void i915_memcpy_init_early(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * Some hypervisors (e.g. KVM) don't support VEX-prefix instructions
> > diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.c b/drivers/gpu/drm/i915/i915_oa_bdw.c
> > index 4acdb94555b7..7c97cd90e973 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_bdw.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_bdw.c
> > @@ -64,28 +64,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_bdw(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"d6de6f55-e526-4f79-a6a6-d7315c09044e",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.h b/drivers/gpu/drm/i915/i915_oa_bdw.h
> > index 0e667f1a8aa1..69479e5b436f 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_bdw.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_bdw.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_BDW_H__
> >   #define __I915_OA_BDW_H__
> > -extern void i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_bdw(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.c b/drivers/gpu/drm/i915/i915_oa_bxt.c
> > index a44195c39923..5ec236b33ecc 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_bxt.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_bxt.c
> > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_bxt(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"5ee72f5c-092f-421e-8b70-225f7c3e9612",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.h b/drivers/gpu/drm/i915/i915_oa_bxt.h
> > index 679e92cf4f1d..47d8327f38a1 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_bxt.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_bxt.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_BXT_H__
> >   #define __I915_OA_BXT_H__
> > -extern void i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_bxt(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.c b/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> > index 7f60d51b8761..3972cd7dcbfc 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"74fb4902-d3d3-4237-9e90-cbdc68d0a446",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.h b/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> > index 4d6025559bbe..434545c9c5d9 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_CFLGT2_H__
> >   #define __I915_OA_CFLGT2_H__
> > -extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.c b/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> > index a92c38e3a0ce..c00b9d9d6ddf 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"577e8e2c-3fa0-4875-8743-3538d585e3b0",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.h b/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> > index 0697f4077402..2c23eda056aa 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_CFLGT3_H__
> >   #define __I915_OA_CFLGT3_H__
> > -extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_chv.c b/drivers/gpu/drm/i915/i915_oa_chv.c
> > index 71ec889a0114..5829d24d0f7c 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_chv.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_chv.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_chv(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"4a534b07-cba3-414d-8d60-874830e883aa",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_chv.h b/drivers/gpu/drm/i915/i915_oa_chv.h
> > index 0986eae3135f..4bb3b6be7bd2 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_chv.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_chv.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_CHV_H__
> >   #define __I915_OA_CHV_H__
> > -extern void i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_chv(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.c b/drivers/gpu/drm/i915/i915_oa_cnl.c
> > index 5c23d883d6c9..fcf08b9e7c73 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cnl.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_cnl.c
> > @@ -75,28 +75,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_cnl(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"db41edd4-d8e7-4730-ad11-b9a2d6833503",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.h b/drivers/gpu/drm/i915/i915_oa_cnl.h
> > index e830a406aff2..c70101655049 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_cnl.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_cnl.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_CNL_H__
> >   #define __I915_OA_CNL_H__
> > -extern void i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_cnl(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_glk.c b/drivers/gpu/drm/i915/i915_oa_glk.c
> > index 4bdda66df7d2..fdb38b9ea9bb 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_glk.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_glk.c
> > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_glk(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"dd3fd789-e783-4204-8cd0-b671bbccb0cf",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_glk.h b/drivers/gpu/drm/i915/i915_oa_glk.h
> > index 06dedf991edb..5d07986827c8 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_glk.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_glk.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_GLK_H__
> >   #define __I915_OA_GLK_H__
> > -extern void i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_glk(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.c b/drivers/gpu/drm/i915/i915_oa_hsw.c
> > index cc6526fdd2bd..640301c9fb9d 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_hsw.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_hsw.c
> > @@ -92,28 +92,28 @@ show_render_basic_id(struct device *kdev, struct device_attribute *attr, char *b
> >   }
> >   void
> > -i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_hsw(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"403d8832-1a27-4aa6-a64e-f5389ce7b212",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_render_basic;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_render_basic;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_render_basic;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_render_basic;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.h b/drivers/gpu/drm/i915/i915_oa_hsw.h
> > index 3d0c870cd0bd..7812a9df8647 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_hsw.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_hsw.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_HSW_H__
> >   #define __I915_OA_HSW_H__
> > -extern void i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_hsw(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_icl.c b/drivers/gpu/drm/i915/i915_oa_icl.c
> > index baa51427a543..2a8c6de619ba 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_icl.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_icl.c
> > @@ -72,28 +72,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_icl(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"a291665e-244b-4b76-9b9a-01de9d3c8068",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_icl.h b/drivers/gpu/drm/i915/i915_oa_icl.h
> > index 24eaa97d61ba..a7c50d31e678 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_icl.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_icl.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_ICL_H__
> >   #define __I915_OA_ICL_H__
> > -extern void i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_icl(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.c b/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> > index 168e49ab0d4d..975311df3d62 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"baa3c7e4-52b6-4b85-801e-465a94b746dd",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.h b/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> > index a55398a904de..53655de7b976 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_KBLGT2_H__
> >   #define __I915_OA_KBLGT2_H__
> > -extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.c b/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> > index 6ffa553c388e..ed39f8dd9af2 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"f1792f32-6db2-4b50-b4b2-557128f1688d",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.h b/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> > index 3ddd3483b7cc..60ebe2b20ada 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_KBLGT3_H__
> >   #define __I915_OA_KBLGT3_H__
> > -extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.c b/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> > index 7ce6ee851d43..60b4dbcd4c59 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.c
> > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"1651949f-0ac0-4cb1-a06f-dafd74a407d1",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.h b/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> > index be6256037239..ee6f747502ea 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_SKLGT2_H__
> >   #define __I915_OA_SKLGT2_H__
> > -extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.c b/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> > index 086ca2631e1c..8c23cf712a7a 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"2b985803-d3c9-4629-8a4f-634bfecba0e8",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.h b/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> > index 650beb068e56..35ccb4125d71 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_SKLGT3_H__
> >   #define __I915_OA_SKLGT3_H__
> > -extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.c b/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> > index b291a6eb8a87..b846da6198f8 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.c
> > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf)
> >   }
> >   void
> > -i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv)
> > +i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915)
> >   {
> > -	strlcpy(dev_priv->perf.oa.test_config.uuid,
> > +	strlcpy(i915->perf.oa.test_config.uuid,
> >   		"882fa433-1f4a-4a67-a962-c741888fe5f5",
> > -		sizeof(dev_priv->perf.oa.test_config.uuid));
> > -	dev_priv->perf.oa.test_config.id = 1;
> > +		sizeof(i915->perf.oa.test_config.uuid));
> > +	i915->perf.oa.test_config.id = 1;
> > -	dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > -	dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > +	i915->perf.oa.test_config.mux_regs = mux_config_test_oa;
> > +	i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa);
> > -	dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > -	dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > +	i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa;
> > +	i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa);
> > -	dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > -	dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > +	i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa;
> > +	i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa);
> > -	dev_priv->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5";
> > -	dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs;
> > +	i915->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5";
> > +	i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs;
> > -	dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr;
> > +	i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > -	dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id";
> > +	i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444;
> > +	i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.h b/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> > index 8dcf849d131e..f3f6d7bf28e0 100644
> > --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.h
> > @@ -10,6 +10,6 @@
> >   #ifndef __I915_OA_SKLGT4_H__
> >   #define __I915_OA_SKLGT4_H__
> > -extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv);
> > +extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915);
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c
> > index d92ddfada262..e0ef77ec4da3 100644
> > --- a/drivers/gpu/drm/i915/i915_perf.c
> > +++ b/drivers/gpu/drm/i915/i915_perf.c
> > @@ -366,7 +366,7 @@ struct perf_open_properties {
> >   	int oa_period_exponent;
> >   };
> > -static void free_oa_config(struct drm_i915_private *dev_priv,
> > +static void free_oa_config(struct drm_i915_private *i915,
> >   			   struct i915_oa_config *oa_config)
> >   {
> >   	if (!PTR_ERR(oa_config->flex_regs))
> > @@ -378,48 +378,48 @@ static void free_oa_config(struct drm_i915_private *dev_priv,
> >   	kfree(oa_config);
> >   }
> > -static void put_oa_config(struct drm_i915_private *dev_priv,
> > +static void put_oa_config(struct drm_i915_private *i915,
> >   			  struct i915_oa_config *oa_config)
> >   {
> >   	if (!atomic_dec_and_test(&oa_config->ref_count))
> >   		return;
> > -	free_oa_config(dev_priv, oa_config);
> > +	free_oa_config(i915, oa_config);
> >   }
> > -static int get_oa_config(struct drm_i915_private *dev_priv,
> > +static int get_oa_config(struct drm_i915_private *i915,
> >   			 int metrics_set,
> >   			 struct i915_oa_config **out_config)
> >   {
> >   	int ret;
> >   	if (metrics_set == 1) {
> > -		*out_config = &dev_priv->perf.oa.test_config;
> > -		atomic_inc(&dev_priv->perf.oa.test_config.ref_count);
> > +		*out_config = &i915->perf.oa.test_config;
> > +		atomic_inc(&i915->perf.oa.test_config.ref_count);
> >   		return 0;
> >   	}
> > -	ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> > +	ret = mutex_lock_interruptible(&i915->perf.metrics_lock);
> >   	if (ret)
> >   		return ret;
> > -	*out_config = idr_find(&dev_priv->perf.metrics_idr, metrics_set);
> > +	*out_config = idr_find(&i915->perf.metrics_idr, metrics_set);
> >   	if (!*out_config)
> >   		ret = -EINVAL;
> >   	else
> >   		atomic_inc(&(*out_config)->ref_count);
> > -	mutex_unlock(&dev_priv->perf.metrics_lock);
> > +	mutex_unlock(&i915->perf.metrics_lock);
> >   	return ret;
> >   }
> > -static u32 gen8_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> > +static u32 gen8_oa_hw_tail_read(struct drm_i915_private *i915)
> >   {
> >   	return I915_READ(GEN8_OATAILPTR) & GEN8_OATAILPTR_MASK;
> >   }
> > -static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> > +static u32 gen7_oa_hw_tail_read(struct drm_i915_private *i915)
> >   {
> >   	u32 oastatus1 = I915_READ(GEN7_OASTATUS1);
> > @@ -428,7 +428,7 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> >   /**
> >    * oa_buffer_check_unlocked - check for data and update tail ptr state
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This is either called via fops (for blocking reads in user ctx) or the poll
> >    * check hrtimer (atomic ctx) to check the OA buffer tail pointer and check
> > @@ -450,9 +450,9 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv)
> >    *
> >    * Returns: %true if the OA buffer contains data, else %false
> >    */
> > -static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> > +static bool oa_buffer_check_unlocked(struct drm_i915_private *i915)
> >   {
> > -	int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> > +	int report_size = i915->perf.oa.oa_buffer.format_size;
> >   	unsigned long flags;
> >   	unsigned int aged_idx;
> >   	u32 head, hw_tail, aged_tail, aging_tail;
> > @@ -462,19 +462,19 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> >   	 * could result in an OA buffer reset which might reset the head,
> >   	 * tails[] and aged_tail state.
> >   	 */
> > -	spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/* NB: The head we observe here might effectively be a little out of
> >   	 * date (between head and tails[aged_idx].offset if there is currently
> >   	 * a read() in progress.
> >   	 */
> > -	head = dev_priv->perf.oa.oa_buffer.head;
> > +	head = i915->perf.oa.oa_buffer.head;
> > -	aged_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> > -	aged_tail = dev_priv->perf.oa.oa_buffer.tails[aged_idx].offset;
> > -	aging_tail = dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset;
> > +	aged_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> > +	aged_tail = i915->perf.oa.oa_buffer.tails[aged_idx].offset;
> > +	aging_tail = i915->perf.oa.oa_buffer.tails[!aged_idx].offset;
> > -	hw_tail = dev_priv->perf.oa.ops.oa_hw_tail_read(dev_priv);
> > +	hw_tail = i915->perf.oa.ops.oa_hw_tail_read(i915);
> >   	/* The tail pointer increases in 64 byte increments,
> >   	 * not in report_size steps...
> > @@ -494,16 +494,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> >   	 * available) without needing to wait for a later hrtimer callback.
> >   	 */
> >   	if (aging_tail != INVALID_TAIL_PTR &&
> > -	    ((now - dev_priv->perf.oa.oa_buffer.aging_timestamp) >
> > +	    ((now - i915->perf.oa.oa_buffer.aging_timestamp) >
> >   	     OA_TAIL_MARGIN_NSEC)) {
> >   		aged_idx ^= 1;
> > -		dev_priv->perf.oa.oa_buffer.aged_tail_idx = aged_idx;
> > +		i915->perf.oa.oa_buffer.aged_tail_idx = aged_idx;
> >   		aged_tail = aging_tail;
> >   		/* Mark that we need a new pointer to start aging... */
> > -		dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR;
> > +		i915->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR;
> >   		aging_tail = INVALID_TAIL_PTR;
> >   	}
> > @@ -518,7 +518,7 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> >   	if (aging_tail == INVALID_TAIL_PTR &&
> >   	    (aged_tail == INVALID_TAIL_PTR ||
> >   	     OA_TAKEN(hw_tail, aged_tail) >= report_size)) {
> > -		struct i915_vma *vma = dev_priv->perf.oa.oa_buffer.vma;
> > +		struct i915_vma *vma = i915->perf.oa.oa_buffer.vma;
> >   		u32 gtt_offset = i915_ggtt_offset(vma);
> >   		/* Be paranoid and do a bounds check on the pointer read back
> > @@ -527,16 +527,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv)
> >   		 */
> >   		if (hw_tail >= gtt_offset &&
> >   		    hw_tail < (gtt_offset + OA_BUFFER_SIZE)) {
> > -			dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset =
> > +			i915->perf.oa.oa_buffer.tails[!aged_idx].offset =
> >   				aging_tail = hw_tail;
> > -			dev_priv->perf.oa.oa_buffer.aging_timestamp = now;
> > +			i915->perf.oa.oa_buffer.aging_timestamp = now;
> >   		} else {
> >   			DRM_ERROR("Ignoring spurious out of range OA buffer tail pointer = %u\n",
> >   				  hw_tail);
> >   		}
> >   	}
> > -	spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	return aged_tail == INVALID_TAIL_PTR ?
> >   		false : OA_TAKEN(aged_tail, head) >= report_size;
> > @@ -599,8 +599,8 @@ static int append_oa_sample(struct i915_perf_stream *stream,
> >   			    size_t *offset,
> >   			    const u8 *report)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > -	int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> > +	struct drm_i915_private *i915 = stream->i915;
> > +	int report_size = i915->perf.oa.oa_buffer.format_size;
> >   	struct drm_i915_perf_record_header header;
> >   	u32 sample_flags = stream->sample_flags;
> > @@ -651,10 +651,10 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   				  size_t count,
> >   				  size_t *offset)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > -	int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> > -	u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr;
> > -	u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> > +	struct drm_i915_private *i915 = stream->i915;
> > +	int report_size = i915->perf.oa.oa_buffer.format_size;
> > +	u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr;
> > +	u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> >   	u32 mask = (OA_BUFFER_SIZE - 1);
> >   	size_t start_offset = *offset;
> >   	unsigned long flags;
> > @@ -666,13 +666,13 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   	if (WARN_ON(!stream->enabled))
> >   		return -EIO;
> > -	spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> > -	head = dev_priv->perf.oa.oa_buffer.head;
> > -	aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> > -	tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> > +	head = i915->perf.oa.oa_buffer.head;
> > +	aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> > +	tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> > -	spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/*
> >   	 * An invalid tail pointer here means we're still waiting for the poll
> > @@ -736,12 +736,12 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   		reason = ((report32[0] >> OAREPORT_REASON_SHIFT) &
> >   			  OAREPORT_REASON_MASK);
> >   		if (reason == 0) {
> > -			if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs))
> > +			if (__ratelimit(&i915->perf.oa.spurious_report_rs))
> >   				DRM_NOTE("Skipping spurious, invalid OA report\n");
> >   			continue;
> >   		}
> > -		ctx_id = report32[2] & dev_priv->perf.oa.specific_ctx_id_mask;
> > +		ctx_id = report32[2] & i915->perf.oa.specific_ctx_id_mask;
> >   		/*
> >   		 * Squash whatever is in the CTX_ID field if it's marked as
> > @@ -751,7 +751,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   		 * Note: that we don't clear the valid_ctx_bit so userspace can
> >   		 * understand that the ID has been squashed by the kernel.
> >   		 */
> > -		if (!(report32[0] & dev_priv->perf.oa.gen8_valid_ctx_bit))
> > +		if (!(report32[0] & i915->perf.oa.gen8_valid_ctx_bit))
> >   			ctx_id = report32[2] = INVALID_CTX_ID;
> >   		/*
> > @@ -785,18 +785,18 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   		 * switches since it's not-uncommon for periodic samples to
> >   		 * identify a switch before any 'context switch' report.
> >   		 */
> > -		if (!dev_priv->perf.oa.exclusive_stream->ctx ||
> > -		    dev_priv->perf.oa.specific_ctx_id == ctx_id ||
> > -		    (dev_priv->perf.oa.oa_buffer.last_ctx_id ==
> > -		     dev_priv->perf.oa.specific_ctx_id) ||
> > +		if (!i915->perf.oa.exclusive_stream->ctx ||
> > +		    i915->perf.oa.specific_ctx_id == ctx_id ||
> > +		    (i915->perf.oa.oa_buffer.last_ctx_id ==
> > +		     i915->perf.oa.specific_ctx_id) ||
> >   		    reason & OAREPORT_REASON_CTX_SWITCH) {
> >   			/*
> >   			 * While filtering for a single context we avoid
> >   			 * leaking the IDs of other contexts.
> >   			 */
> > -			if (dev_priv->perf.oa.exclusive_stream->ctx &&
> > -			    dev_priv->perf.oa.specific_ctx_id != ctx_id) {
> > +			if (i915->perf.oa.exclusive_stream->ctx &&
> > +			    i915->perf.oa.specific_ctx_id != ctx_id) {
> >   				report32[2] = INVALID_CTX_ID;
> >   			}
> > @@ -805,7 +805,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   			if (ret)
> >   				break;
> > -			dev_priv->perf.oa.oa_buffer.last_ctx_id = ctx_id;
> > +			i915->perf.oa.oa_buffer.last_ctx_id = ctx_id;
> >   		}
> >   		/*
> > @@ -819,7 +819,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   	}
> >   	if (start_offset != *offset) {
> > -		spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +		spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   		/*
> >   		 * We removed the gtt_offset for the copy loop above, indexing
> > @@ -828,9 +828,9 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream,
> >   		head += gtt_offset;
> >   		I915_WRITE(GEN8_OAHEADPTR, head & GEN8_OAHEADPTR_MASK);
> > -		dev_priv->perf.oa.oa_buffer.head = head;
> > +		i915->perf.oa.oa_buffer.head = head;
> > -		spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +		spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	}
> >   	return ret;
> > @@ -861,11 +861,11 @@ static int gen8_oa_read(struct i915_perf_stream *stream,
> >   			size_t count,
> >   			size_t *offset)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	u32 oastatus;
> >   	int ret;
> > -	if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr))
> > +	if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr))
> >   		return -EIO;
> >   	oastatus = I915_READ(GEN8_OASTATUS);
> > @@ -891,10 +891,10 @@ static int gen8_oa_read(struct i915_perf_stream *stream,
> >   			return ret;
> >   		DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n",
> > -			  dev_priv->perf.oa.period_exponent);
> > +			  i915->perf.oa.period_exponent);
> > -		dev_priv->perf.oa.ops.oa_disable(stream);
> > -		dev_priv->perf.oa.ops.oa_enable(stream);
> > +		i915->perf.oa.ops.oa_disable(stream);
> > +		i915->perf.oa.ops.oa_enable(stream);
> >   		/*
> >   		 * Note: .oa_enable() is expected to re-init the oabuffer and
> > @@ -940,10 +940,10 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> >   				  size_t count,
> >   				  size_t *offset)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > -	int report_size = dev_priv->perf.oa.oa_buffer.format_size;
> > -	u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr;
> > -	u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> > +	struct drm_i915_private *i915 = stream->i915;
> > +	int report_size = i915->perf.oa.oa_buffer.format_size;
> > +	u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr;
> > +	u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> >   	u32 mask = (OA_BUFFER_SIZE - 1);
> >   	size_t start_offset = *offset;
> >   	unsigned long flags;
> > @@ -955,13 +955,13 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> >   	if (WARN_ON(!stream->enabled))
> >   		return -EIO;
> > -	spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> > -	head = dev_priv->perf.oa.oa_buffer.head;
> > -	aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx;
> > -	tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> > +	head = i915->perf.oa.oa_buffer.head;
> > +	aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx;
> > +	tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset;
> > -	spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/* An invalid tail pointer here means we're still waiting for the poll
> >   	 * hrtimer callback to give us a pointer
> > @@ -1014,7 +1014,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> >   		 * copying it to userspace...
> >   		 */
> >   		if (report32[0] == 0) {
> > -			if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs))
> > +			if (__ratelimit(&i915->perf.oa.spurious_report_rs))
> >   				DRM_NOTE("Skipping spurious, invalid OA report\n");
> >   			continue;
> >   		}
> > @@ -1033,7 +1033,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> >   	}
> >   	if (start_offset != *offset) {
> > -		spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +		spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   		/* We removed the gtt_offset for the copy loop above, indexing
> >   		 * relative to oa_buf_base so put back here...
> > @@ -1043,9 +1043,9 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream,
> >   		I915_WRITE(GEN7_OASTATUS2,
> >   			   ((head & GEN7_OASTATUS2_HEAD_MASK) |
> >   			    GEN7_OASTATUS2_MEM_SELECT_GGTT));
> > -		dev_priv->perf.oa.oa_buffer.head = head;
> > +		i915->perf.oa.oa_buffer.head = head;
> > -		spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +		spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	}
> >   	return ret;
> > @@ -1072,11 +1072,11 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> >   			size_t count,
> >   			size_t *offset)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	u32 oastatus1;
> >   	int ret;
> > -	if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr))
> > +	if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr))
> >   		return -EIO;
> >   	oastatus1 = I915_READ(GEN7_OASTATUS1);
> > @@ -1086,7 +1086,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> >   	 * may be updated asynchronously) so we ignore status bits
> >   	 * that have already been reported to userspace.
> >   	 */
> > -	oastatus1 &= ~dev_priv->perf.oa.gen7_latched_oastatus1;
> > +	oastatus1 &= ~i915->perf.oa.gen7_latched_oastatus1;
> >   	/* We treat OABUFFER_OVERFLOW as a significant error:
> >   	 *
> > @@ -1115,10 +1115,10 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> >   			return ret;
> >   		DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n",
> > -			  dev_priv->perf.oa.period_exponent);
> > +			  i915->perf.oa.period_exponent);
> > -		dev_priv->perf.oa.ops.oa_disable(stream);
> > -		dev_priv->perf.oa.ops.oa_enable(stream);
> > +		i915->perf.oa.ops.oa_disable(stream);
> > +		i915->perf.oa.ops.oa_enable(stream);
> >   		oastatus1 = I915_READ(GEN7_OASTATUS1);
> >   	}
> > @@ -1128,7 +1128,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> >   				       DRM_I915_PERF_RECORD_OA_REPORT_LOST);
> >   		if (ret)
> >   			return ret;
> > -		dev_priv->perf.oa.gen7_latched_oastatus1 |=
> > +		i915->perf.oa.gen7_latched_oastatus1 |=
> >   			GEN7_OASTATUS1_REPORT_LOST;
> >   	}
> > @@ -1151,14 +1151,14 @@ static int gen7_oa_read(struct i915_perf_stream *stream,
> >    */
> >   static int i915_oa_wait_unlocked(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	/* We would wait indefinitely if periodic sampling is not enabled */
> > -	if (!dev_priv->perf.oa.periodic)
> > +	if (!i915->perf.oa.periodic)
> >   		return -EIO;
> > -	return wait_event_interruptible(dev_priv->perf.oa.poll_wq,
> > -					oa_buffer_check_unlocked(dev_priv));
> > +	return wait_event_interruptible(i915->perf.oa.poll_wq,
> > +					oa_buffer_check_unlocked(i915));
> >   }
> >   /**
> > @@ -1175,9 +1175,9 @@ static void i915_oa_poll_wait(struct i915_perf_stream *stream,
> >   			      struct file *file,
> >   			      poll_table *wait)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	poll_wait(file, &dev_priv->perf.oa.poll_wq, wait);
> > +	poll_wait(file, &i915->perf.oa.poll_wq, wait);
> >   }
> >   /**
> > @@ -1197,9 +1197,9 @@ static int i915_oa_read(struct i915_perf_stream *stream,
> >   			size_t count,
> >   			size_t *offset)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	return dev_priv->perf.oa.ops.read(stream, buf, count, offset);
> > +	return i915->perf.oa.ops.read(stream, buf, count, offset);
> >   }
> >   static struct intel_context *oa_pin_context(struct drm_i915_private *i915,
> > @@ -1248,7 +1248,7 @@ static struct intel_context *oa_pin_context(struct drm_i915_private *i915,
> >    */
> >   static int oa_get_render_ctx_id(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *i915 = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	struct intel_context *ce;
> >   	ce = oa_pin_context(i915, stream->ctx);
> > @@ -1330,17 +1330,17 @@ static int oa_get_render_ctx_id(struct i915_perf_stream *stream)
> >    */
> >   static void oa_put_render_ctx_id(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	struct intel_context *ce;
> > -	dev_priv->perf.oa.specific_ctx_id = INVALID_CTX_ID;
> > -	dev_priv->perf.oa.specific_ctx_id_mask = 0;
> > +	i915->perf.oa.specific_ctx_id = INVALID_CTX_ID;
> > +	i915->perf.oa.specific_ctx_id_mask = 0;
> > -	ce = fetch_and_zero(&dev_priv->perf.oa.pinned_ctx);
> > +	ce = fetch_and_zero(&i915->perf.oa.pinned_ctx);
> >   	if (ce) {
> > -		mutex_lock(&dev_priv->drm.struct_mutex);
> > +		mutex_lock(&i915->drm.struct_mutex);
> >   		intel_context_unpin(ce);
> > -		mutex_unlock(&dev_priv->drm.struct_mutex);
> > +		mutex_unlock(&i915->drm.struct_mutex);
> >   	}
> >   }
> > @@ -1359,64 +1359,64 @@ free_oa_buffer(struct drm_i915_private *i915)
> >   static void i915_oa_stream_destroy(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	BUG_ON(stream != dev_priv->perf.oa.exclusive_stream);
> > +	BUG_ON(stream != i915->perf.oa.exclusive_stream);
> >   	/*
> >   	 * Unset exclusive_stream first, it will be checked while disabling
> >   	 * the metric set on gen8+.
> >   	 */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	dev_priv->perf.oa.exclusive_stream = NULL;
> > -	dev_priv->perf.oa.ops.disable_metric_set(dev_priv);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	i915->perf.oa.exclusive_stream = NULL;
> > +	i915->perf.oa.ops.disable_metric_set(i915);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	free_oa_buffer(dev_priv);
> > +	free_oa_buffer(i915);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > -	intel_runtime_pm_put(dev_priv, stream->wakeref);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > +	intel_runtime_pm_put(i915, stream->wakeref);
> >   	if (stream->ctx)
> >   		oa_put_render_ctx_id(stream);
> > -	put_oa_config(dev_priv, stream->oa_config);
> > +	put_oa_config(i915, stream->oa_config);
> > -	if (dev_priv->perf.oa.spurious_report_rs.missed) {
> > +	if (i915->perf.oa.spurious_report_rs.missed) {
> >   		DRM_NOTE("%d spurious OA report notices suppressed due to ratelimiting\n",
> > -			 dev_priv->perf.oa.spurious_report_rs.missed);
> > +			 i915->perf.oa.spurious_report_rs.missed);
> >   	}
> >   }
> > -static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv)
> > +static void gen7_init_oa_buffer(struct drm_i915_private *i915)
> >   {
> > -	u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> > +	u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> >   	unsigned long flags;
> > -	spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/* Pre-DevBDW: OABUFFER must be set with counters off,
> >   	 * before OASTATUS1, but after OASTATUS2
> >   	 */
> >   	I915_WRITE(GEN7_OASTATUS2,
> >   		   gtt_offset | GEN7_OASTATUS2_MEM_SELECT_GGTT); /* head */
> > -	dev_priv->perf.oa.oa_buffer.head = gtt_offset;
> > +	i915->perf.oa.oa_buffer.head = gtt_offset;
> >   	I915_WRITE(GEN7_OABUFFER, gtt_offset);
> >   	I915_WRITE(GEN7_OASTATUS1, gtt_offset | OABUFFER_SIZE_16M); /* tail */
> >   	/* Mark that we need updated tail pointers to read from... */
> > -	dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> > -	dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> > +	i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> > +	i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> > -	spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/* On Haswell we have to track which OASTATUS1 flags we've
> >   	 * already seen since they can't be cleared while periodic
> >   	 * sampling is enabled.
> >   	 */
> > -	dev_priv->perf.oa.gen7_latched_oastatus1 = 0;
> > +	i915->perf.oa.gen7_latched_oastatus1 = 0;
> >   	/* NB: although the OA buffer will initially be allocated
> >   	 * zeroed via shmfs (and so this memset is redundant when
> > @@ -1429,24 +1429,24 @@ static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv)
> >   	 * the assumption that new reports are being written to zeroed
> >   	 * memory...
> >   	 */
> > -	memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> > +	memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> >   	/* Maybe make ->pollin per-stream state if we support multiple
> >   	 * concurrent streams in the future.
> >   	 */
> > -	dev_priv->perf.oa.pollin = false;
> > +	i915->perf.oa.pollin = false;
> >   }
> > -static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> > +static void gen8_init_oa_buffer(struct drm_i915_private *i915)
> >   {
> > -	u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma);
> > +	u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma);
> >   	unsigned long flags;
> > -	spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	I915_WRITE(GEN8_OASTATUS, 0);
> >   	I915_WRITE(GEN8_OAHEADPTR, gtt_offset);
> > -	dev_priv->perf.oa.oa_buffer.head = gtt_offset;
> > +	i915->perf.oa.oa_buffer.head = gtt_offset;
> >   	I915_WRITE(GEN8_OABUFFER_UDW, 0);
> > @@ -1463,17 +1463,17 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN8_OATAILPTR, gtt_offset & GEN8_OATAILPTR_MASK);
> >   	/* Mark that we need updated tail pointers to read from... */
> > -	dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> > -	dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> > +	i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR;
> > +	i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR;
> >   	/*
> >   	 * Reset state used to recognise context switches, affecting which
> >   	 * reports we will forward to userspace while filtering for a single
> >   	 * context.
> >   	 */
> > -	dev_priv->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID;
> > +	i915->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID;
> > -	spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags);
> > +	spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags);
> >   	/*
> >   	 * NB: although the OA buffer will initially be allocated
> > @@ -1487,32 +1487,32 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv)
> >   	 * the assumption that new reports are being written to zeroed
> >   	 * memory...
> >   	 */
> > -	memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> > +	memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE);
> >   	/*
> >   	 * Maybe make ->pollin per-stream state if we support multiple
> >   	 * concurrent streams in the future.
> >   	 */
> > -	dev_priv->perf.oa.pollin = false;
> > +	i915->perf.oa.pollin = false;
> >   }
> > -static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> > +static int alloc_oa_buffer(struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_gem_object *bo;
> >   	struct i915_vma *vma;
> >   	int ret;
> > -	if (WARN_ON(dev_priv->perf.oa.oa_buffer.vma))
> > +	if (WARN_ON(i915->perf.oa.oa_buffer.vma))
> >   		return -ENODEV;
> > -	ret = i915_mutex_lock_interruptible(&dev_priv->drm);
> > +	ret = i915_mutex_lock_interruptible(&i915->drm);
> >   	if (ret)
> >   		return ret;
> >   	BUILD_BUG_ON_NOT_POWER_OF_2(OA_BUFFER_SIZE);
> >   	BUILD_BUG_ON(OA_BUFFER_SIZE < SZ_128K || OA_BUFFER_SIZE > SZ_16M);
> > -	bo = i915_gem_object_create_shmem(dev_priv, OA_BUFFER_SIZE);
> > +	bo = i915_gem_object_create_shmem(i915, OA_BUFFER_SIZE);
> >   	if (IS_ERR(bo)) {
> >   		DRM_ERROR("Failed to allocate OA buffer\n");
> >   		ret = PTR_ERR(bo);
> > @@ -1527,18 +1527,18 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> >   		ret = PTR_ERR(vma);
> >   		goto err_unref;
> >   	}
> > -	dev_priv->perf.oa.oa_buffer.vma = vma;
> > +	i915->perf.oa.oa_buffer.vma = vma;
> > -	dev_priv->perf.oa.oa_buffer.vaddr =
> > +	i915->perf.oa.oa_buffer.vaddr =
> >   		i915_gem_object_pin_map(bo, I915_MAP_WB);
> > -	if (IS_ERR(dev_priv->perf.oa.oa_buffer.vaddr)) {
> > -		ret = PTR_ERR(dev_priv->perf.oa.oa_buffer.vaddr);
> > +	if (IS_ERR(i915->perf.oa.oa_buffer.vaddr)) {
> > +		ret = PTR_ERR(i915->perf.oa.oa_buffer.vaddr);
> >   		goto err_unpin;
> >   	}
> >   	DRM_DEBUG_DRIVER("OA Buffer initialized, gtt offset = 0x%x, vaddr = %p\n",
> > -			 i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma),
> > -			 dev_priv->perf.oa.oa_buffer.vaddr);
> > +			 i915_ggtt_offset(i915->perf.oa.oa_buffer.vma),
> > +			 i915->perf.oa.oa_buffer.vaddr);
> >   	goto unlock;
> > @@ -1548,15 +1548,15 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv)
> >   err_unref:
> >   	i915_gem_object_put(bo);
> > -	dev_priv->perf.oa.oa_buffer.vaddr = NULL;
> > -	dev_priv->perf.oa.oa_buffer.vma = NULL;
> > +	i915->perf.oa.oa_buffer.vaddr = NULL;
> > +	i915->perf.oa.oa_buffer.vma = NULL;
> >   unlock:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return ret;
> >   }
> > -static void config_oa_regs(struct drm_i915_private *dev_priv,
> > +static void config_oa_regs(struct drm_i915_private *i915,
> >   			   const struct i915_oa_reg *regs,
> >   			   u32 n_regs)
> >   {
> > @@ -1571,7 +1571,7 @@ static void config_oa_regs(struct drm_i915_private *dev_priv,
> >   static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	const struct i915_oa_config *oa_config = stream->oa_config;
> >   	/* PRM:
> > @@ -1588,7 +1588,7 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> >   	I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) |
> >   				  GEN6_CSUNIT_CLOCK_GATE_DISABLE));
> > -	config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len);
> > +	config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len);
> >   	/* It apparently takes a fairly long time for a new MUX
> >   	 * configuration to be be applied after these register writes.
> > @@ -1613,13 +1613,13 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream)
> >   	 */
> >   	usleep_range(15000, 20000);
> > -	config_oa_regs(dev_priv, oa_config->b_counter_regs,
> > +	config_oa_regs(i915, oa_config->b_counter_regs,
> >   		       oa_config->b_counter_regs_len);
> >   	return 0;
> >   }
> > -static void hsw_disable_metric_set(struct drm_i915_private *dev_priv)
> > +static void hsw_disable_metric_set(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) &
> >   				  ~GEN6_CSUNIT_CLOCK_GATE_DISABLE));
> > @@ -1716,15 +1716,15 @@ gen8_update_reg_state_unlocked(struct intel_context *ce,
> >    *
> >    * Note: it's only the RCS/Render context that has any OA state.
> >    */
> > -static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> > +static int gen8_configure_all_contexts(struct drm_i915_private *i915,
> >   				       const struct i915_oa_config *oa_config)
> >   {
> > -	unsigned int map_type = i915_coherent_map_type(dev_priv);
> > +	unsigned int map_type = i915_coherent_map_type(i915);
> >   	struct i915_gem_context *ctx;
> >   	struct i915_request *rq;
> >   	int ret;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	/*
> >   	 * The OA register config is setup through the context image. This image
> > @@ -1739,14 +1739,14 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> >   	 * So far the best way to work around this issue seems to be draining
> >   	 * the GPU from any submitted work.
> >   	 */
> > -	ret = i915_gem_wait_for_idle(dev_priv,
> > +	ret = i915_gem_wait_for_idle(i915,
> >   				     I915_WAIT_LOCKED,
> >   				     MAX_SCHEDULE_TIMEOUT);
> >   	if (ret)
> >   		return ret;
> >   	/* Update all contexts now that we've stalled the submission. */
> > -	list_for_each_entry(ctx, &dev_priv->contexts.list, link) {
> > +	list_for_each_entry(ctx, &i915->contexts.list, link) {
> >   		struct i915_gem_engines_iter it;
> >   		struct intel_context *ce;
> > @@ -1783,7 +1783,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> >   	 * Apply the configuration by doing one context restore of the edited
> >   	 * context image.
> >   	 */
> > -	rq = i915_request_create(dev_priv->engine[RCS0]->kernel_context);
> > +	rq = i915_request_create(i915->engine[RCS0]->kernel_context);
> >   	if (IS_ERR(rq))
> >   		return PTR_ERR(rq);
> > @@ -1794,7 +1794,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv,
> >   static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	const struct i915_oa_config *oa_config = stream->oa_config;
> >   	int ret;
> > @@ -1821,7 +1821,7 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> >   	 * be read back from automatically triggered reports, as part of the
> >   	 * RPT_ID field.
> >   	 */
> > -	if (IS_GEN_RANGE(dev_priv, 9, 11)) {
> > +	if (IS_GEN_RANGE(i915, 9, 11)) {
> >   		I915_WRITE(GEN8_OA_DEBUG,
> >   			   _MASKED_BIT_ENABLE(GEN9_OA_DEBUG_DISABLE_CLK_RATIO_REPORTS |
> >   					      GEN9_OA_DEBUG_INCLUDE_CLK_RATIO));
> > @@ -1832,31 +1832,31 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream)
> >   	 * to make sure all slices/subslices are ON before writing to NOA
> >   	 * registers.
> >   	 */
> > -	ret = gen8_configure_all_contexts(dev_priv, oa_config);
> > +	ret = gen8_configure_all_contexts(i915, oa_config);
> >   	if (ret)
> >   		return ret;
> > -	config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len);
> > +	config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len);
> > -	config_oa_regs(dev_priv, oa_config->b_counter_regs,
> > +	config_oa_regs(i915, oa_config->b_counter_regs,
> >   		       oa_config->b_counter_regs_len);
> >   	return 0;
> >   }
> > -static void gen8_disable_metric_set(struct drm_i915_private *dev_priv)
> > +static void gen8_disable_metric_set(struct drm_i915_private *i915)
> >   {
> >   	/* Reset all contexts' slices/subslices configurations. */
> > -	gen8_configure_all_contexts(dev_priv, NULL);
> > +	gen8_configure_all_contexts(i915, NULL);
> >   	I915_WRITE(GDT_CHICKEN_BITS, (I915_READ(GDT_CHICKEN_BITS) &
> >   				      ~GT_NOA_ENABLE));
> >   }
> > -static void gen10_disable_metric_set(struct drm_i915_private *dev_priv)
> > +static void gen10_disable_metric_set(struct drm_i915_private *i915)
> >   {
> >   	/* Reset all contexts' slices/subslices configurations. */
> > -	gen8_configure_all_contexts(dev_priv, NULL);
> > +	gen8_configure_all_contexts(i915, NULL);
> >   	/* Make sure we disable noa to save power. */
> >   	I915_WRITE(RPM_CONFIG1,
> > @@ -1865,12 +1865,12 @@ static void gen10_disable_metric_set(struct drm_i915_private *dev_priv)
> >   static void gen7_oa_enable(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	struct i915_gem_context *ctx = stream->ctx;
> > -	u32 ctx_id = dev_priv->perf.oa.specific_ctx_id;
> > -	bool periodic = dev_priv->perf.oa.periodic;
> > -	u32 period_exponent = dev_priv->perf.oa.period_exponent;
> > -	u32 report_format = dev_priv->perf.oa.oa_buffer.format;
> > +	u32 ctx_id = i915->perf.oa.specific_ctx_id;
> > +	bool periodic = i915->perf.oa.periodic;
> > +	u32 period_exponent = i915->perf.oa.period_exponent;
> > +	u32 report_format = i915->perf.oa.oa_buffer.format;
> >   	/*
> >   	 * Reset buf pointers so we don't forward reports from before now.
> > @@ -1881,7 +1881,7 @@ static void gen7_oa_enable(struct i915_perf_stream *stream)
> >   	 * on the assumption that certain fields are written to zeroed
> >   	 * memory which this helps maintains.
> >   	 */
> > -	gen7_init_oa_buffer(dev_priv);
> > +	gen7_init_oa_buffer(i915);
> >   	I915_WRITE(GEN7_OACONTROL,
> >   		   (ctx_id & GEN7_OACONTROL_CTX_MASK) |
> > @@ -1895,8 +1895,8 @@ static void gen7_oa_enable(struct i915_perf_stream *stream)
> >   static void gen8_oa_enable(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > -	u32 report_format = dev_priv->perf.oa.oa_buffer.format;
> > +	struct drm_i915_private *i915 = stream->i915;
> > +	u32 report_format = i915->perf.oa.oa_buffer.format;
> >   	/*
> >   	 * Reset buf pointers so we don't forward reports from before now.
> > @@ -1907,7 +1907,7 @@ static void gen8_oa_enable(struct i915_perf_stream *stream)
> >   	 * on the assumption that certain fields are written to zeroed
> >   	 * memory which this helps maintains.
> >   	 */
> > -	gen8_init_oa_buffer(dev_priv);
> > +	gen8_init_oa_buffer(i915);
> >   	/*
> >   	 * Note: we don't rely on the hardware to perform single context
> > @@ -1930,19 +1930,19 @@ static void gen8_oa_enable(struct i915_perf_stream *stream)
> >    */
> >   static void i915_oa_stream_enable(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	dev_priv->perf.oa.ops.oa_enable(stream);
> > +	i915->perf.oa.ops.oa_enable(stream);
> > -	if (dev_priv->perf.oa.periodic)
> > -		hrtimer_start(&dev_priv->perf.oa.poll_check_timer,
> > +	if (i915->perf.oa.periodic)
> > +		hrtimer_start(&i915->perf.oa.poll_check_timer,
> >   			      ns_to_ktime(POLL_PERIOD),
> >   			      HRTIMER_MODE_REL_PINNED);
> >   }
> >   static void gen7_oa_disable(struct i915_perf_stream *stream)
> >   {
> > -	struct intel_uncore *uncore = &stream->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &stream->i915->uncore;
> >   	intel_uncore_write(uncore, GEN7_OACONTROL, 0);
> >   	if (intel_wait_for_register(uncore,
> > @@ -1953,7 +1953,7 @@ static void gen7_oa_disable(struct i915_perf_stream *stream)
> >   static void gen8_oa_disable(struct i915_perf_stream *stream)
> >   {
> > -	struct intel_uncore *uncore = &stream->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &stream->i915->uncore;
> >   	intel_uncore_write(uncore, GEN8_OACONTROL, 0);
> >   	if (intel_wait_for_register(uncore,
> > @@ -1972,12 +1972,12 @@ static void gen8_oa_disable(struct i915_perf_stream *stream)
> >    */
> >   static void i915_oa_stream_disable(struct i915_perf_stream *stream)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	dev_priv->perf.oa.ops.oa_disable(stream);
> > +	i915->perf.oa.ops.oa_disable(stream);
> > -	if (dev_priv->perf.oa.periodic)
> > -		hrtimer_cancel(&dev_priv->perf.oa.poll_check_timer);
> > +	if (i915->perf.oa.periodic)
> > +		hrtimer_cancel(&i915->perf.oa.poll_check_timer);
> >   }
> >   static const struct i915_perf_stream_ops i915_oa_stream_ops = {
> > @@ -2011,7 +2011,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   			       struct drm_i915_perf_open_param *param,
> >   			       struct perf_open_properties *props)
> >   {
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	int format_size;
> >   	int ret;
> > @@ -2019,7 +2019,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   	 * reason then don't let userspace try their luck with config
> >   	 * IDs
> >   	 */
> > -	if (!dev_priv->perf.metrics_kobj) {
> > +	if (!i915->perf.metrics_kobj) {
> >   		DRM_DEBUG("OA metrics weren't advertised via sysfs\n");
> >   		return -EINVAL;
> >   	}
> > @@ -2029,7 +2029,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   		return -EINVAL;
> >   	}
> > -	if (!dev_priv->perf.oa.ops.enable_metric_set) {
> > +	if (!i915->perf.oa.ops.enable_metric_set) {
> >   		DRM_DEBUG("OA unit not supported\n");
> >   		return -ENODEV;
> >   	}
> > @@ -2038,7 +2038,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   	 * counter reports and marshal to the appropriate client
> >   	 * we currently only allow exclusive access
> >   	 */
> > -	if (dev_priv->perf.oa.exclusive_stream) {
> > +	if (i915->perf.oa.exclusive_stream) {
> >   		DRM_DEBUG("OA unit already in use\n");
> >   		return -EBUSY;
> >   	}
> > @@ -2059,32 +2059,32 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   	 *
> >   	 * Using the same limiting factors as printk_ratelimit()
> >   	 */
> > -	ratelimit_state_init(&dev_priv->perf.oa.spurious_report_rs,
> > +	ratelimit_state_init(&i915->perf.oa.spurious_report_rs,
> >   			     5 * HZ, 10);
> >   	/* Since we use a DRM_NOTE for spurious reports it would be
> >   	 * inconsistent to let __ratelimit() automatically print a warning for
> >   	 * throttling.
> >   	 */
> > -	ratelimit_set_flags(&dev_priv->perf.oa.spurious_report_rs,
> > +	ratelimit_set_flags(&i915->perf.oa.spurious_report_rs,
> >   			    RATELIMIT_MSG_ON_RELEASE);
> >   	stream->sample_size = sizeof(struct drm_i915_perf_record_header);
> > -	format_size = dev_priv->perf.oa.oa_formats[props->oa_format].size;
> > +	format_size = i915->perf.oa.oa_formats[props->oa_format].size;
> >   	stream->sample_flags |= SAMPLE_OA_REPORT;
> >   	stream->sample_size += format_size;
> > -	dev_priv->perf.oa.oa_buffer.format_size = format_size;
> > -	if (WARN_ON(dev_priv->perf.oa.oa_buffer.format_size == 0))
> > +	i915->perf.oa.oa_buffer.format_size = format_size;
> > +	if (WARN_ON(i915->perf.oa.oa_buffer.format_size == 0))
> >   		return -EINVAL;
> > -	dev_priv->perf.oa.oa_buffer.format =
> > -		dev_priv->perf.oa.oa_formats[props->oa_format].format;
> > +	i915->perf.oa.oa_buffer.format =
> > +		i915->perf.oa.oa_formats[props->oa_format].format;
> > -	dev_priv->perf.oa.periodic = props->oa_periodic;
> > -	if (dev_priv->perf.oa.periodic)
> > -		dev_priv->perf.oa.period_exponent = props->oa_period_exponent;
> > +	i915->perf.oa.periodic = props->oa_periodic;
> > +	if (i915->perf.oa.periodic)
> > +		i915->perf.oa.period_exponent = props->oa_period_exponent;
> >   	if (stream->ctx) {
> >   		ret = oa_get_render_ctx_id(stream);
> > @@ -2094,7 +2094,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   		}
> >   	}
> > -	ret = get_oa_config(dev_priv, props->metrics_set, &stream->oa_config);
> > +	ret = get_oa_config(i915, props->metrics_set, &stream->oa_config);
> >   	if (ret) {
> >   		DRM_DEBUG("Invalid OA config id=%i\n", props->metrics_set);
> >   		goto err_config;
> > @@ -2112,43 +2112,43 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream,
> >   	 *   In our case we are expecting that taking pm + FORCEWAKE
> >   	 *   references will effectively disable RC6.
> >   	 */
> > -	stream->wakeref = intel_runtime_pm_get(dev_priv);
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	stream->wakeref = intel_runtime_pm_get(i915);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> > -	ret = alloc_oa_buffer(dev_priv);
> > +	ret = alloc_oa_buffer(i915);
> >   	if (ret)
> >   		goto err_oa_buf_alloc;
> > -	ret = i915_mutex_lock_interruptible(&dev_priv->drm);
> > +	ret = i915_mutex_lock_interruptible(&i915->drm);
> >   	if (ret)
> >   		goto err_lock;
> >   	stream->ops = &i915_oa_stream_ops;
> > -	dev_priv->perf.oa.exclusive_stream = stream;
> > +	i915->perf.oa.exclusive_stream = stream;
> > -	ret = dev_priv->perf.oa.ops.enable_metric_set(stream);
> > +	ret = i915->perf.oa.ops.enable_metric_set(stream);
> >   	if (ret) {
> >   		DRM_DEBUG("Unable to enable metric set\n");
> >   		goto err_enable;
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return 0;
> >   err_enable:
> > -	dev_priv->perf.oa.exclusive_stream = NULL;
> > -	dev_priv->perf.oa.ops.disable_metric_set(dev_priv);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	i915->perf.oa.exclusive_stream = NULL;
> > +	i915->perf.oa.ops.disable_metric_set(i915);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   err_lock:
> > -	free_oa_buffer(dev_priv);
> > +	free_oa_buffer(i915);
> >   err_oa_buf_alloc:
> > -	put_oa_config(dev_priv, stream->oa_config);
> > +	put_oa_config(i915, stream->oa_config);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > -	intel_runtime_pm_put(dev_priv, stream->wakeref);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > +	intel_runtime_pm_put(i915, stream->wakeref);
> >   err_config:
> >   	if (stream->ctx)
> > @@ -2238,7 +2238,7 @@ static ssize_t i915_perf_read(struct file *file,
> >   			      loff_t *ppos)
> >   {
> >   	struct i915_perf_stream *stream = file->private_data;
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	ssize_t ret;
> >   	/* To ensure it's handled consistently we simply treat all reads of a
> > @@ -2261,15 +2261,15 @@ static ssize_t i915_perf_read(struct file *file,
> >   			if (ret)
> >   				return ret;
> > -			mutex_lock(&dev_priv->perf.lock);
> > +			mutex_lock(&i915->perf.lock);
> >   			ret = i915_perf_read_locked(stream, file,
> >   						    buf, count, ppos);
> > -			mutex_unlock(&dev_priv->perf.lock);
> > +			mutex_unlock(&i915->perf.lock);
> >   		} while (ret == -EAGAIN);
> >   	} else {
> > -		mutex_lock(&dev_priv->perf.lock);
> > +		mutex_lock(&i915->perf.lock);
> >   		ret = i915_perf_read_locked(stream, file, buf, count, ppos);
> > -		mutex_unlock(&dev_priv->perf.lock);
> > +		mutex_unlock(&i915->perf.lock);
> >   	}
> >   	/* We allow the poll checking to sometimes report false positive EPOLLIN
> > @@ -2284,7 +2284,7 @@ static ssize_t i915_perf_read(struct file *file,
> >   		/* Maybe make ->pollin per-stream state if we support multiple
> >   		 * concurrent streams in the future.
> >   		 */
> > -		dev_priv->perf.oa.pollin = false;
> > +		i915->perf.oa.pollin = false;
> >   	}
> >   	return ret;
> > @@ -2292,13 +2292,13 @@ static ssize_t i915_perf_read(struct file *file,
> >   static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(hrtimer, typeof(*dev_priv),
> > +	struct drm_i915_private *i915 =
> > +		container_of(hrtimer, typeof(*i915),
> >   			     perf.oa.poll_check_timer);
> > -	if (oa_buffer_check_unlocked(dev_priv)) {
> > -		dev_priv->perf.oa.pollin = true;
> > -		wake_up(&dev_priv->perf.oa.poll_wq);
> > +	if (oa_buffer_check_unlocked(i915)) {
> > +		i915->perf.oa.pollin = true;
> > +		wake_up(&i915->perf.oa.poll_wq);
> >   	}
> >   	hrtimer_forward_now(hrtimer, ns_to_ktime(POLL_PERIOD));
> > @@ -2308,7 +2308,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
> >   /**
> >    * i915_perf_poll_locked - poll_wait() with a suitable wait queue for stream
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @stream: An i915 perf stream
> >    * @file: An i915 perf stream file
> >    * @wait: poll() state table
> > @@ -2322,7 +2322,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer)
> >    *
> >    * Returns: any poll events that are ready without sleeping
> >    */
> > -static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> > +static __poll_t i915_perf_poll_locked(struct drm_i915_private *i915,
> >   					  struct i915_perf_stream *stream,
> >   					  struct file *file,
> >   					  poll_table *wait)
> > @@ -2337,7 +2337,7 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> >   	 * the hrtimer/oa_poll_check_timer_cb to notify us when there are
> >   	 * samples to read.
> >   	 */
> > -	if (dev_priv->perf.oa.pollin)
> > +	if (i915->perf.oa.pollin)
> >   		events |= EPOLLIN;
> >   	return events;
> > @@ -2359,12 +2359,12 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv,
> >   static __poll_t i915_perf_poll(struct file *file, poll_table *wait)
> >   {
> >   	struct i915_perf_stream *stream = file->private_data;
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	__poll_t ret;
> > -	mutex_lock(&dev_priv->perf.lock);
> > -	ret = i915_perf_poll_locked(dev_priv, stream, file, wait);
> > -	mutex_unlock(&dev_priv->perf.lock);
> > +	mutex_lock(&i915->perf.lock);
> > +	ret = i915_perf_poll_locked(i915, stream, file, wait);
> > +	mutex_unlock(&i915->perf.lock);
> >   	return ret;
> >   }
> > @@ -2461,12 +2461,12 @@ static long i915_perf_ioctl(struct file *file,
> >   			    unsigned long arg)
> >   {
> >   	struct i915_perf_stream *stream = file->private_data;
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> >   	long ret;
> > -	mutex_lock(&dev_priv->perf.lock);
> > +	mutex_lock(&i915->perf.lock);
> >   	ret = i915_perf_ioctl_locked(stream, cmd, arg);
> > -	mutex_unlock(&dev_priv->perf.lock);
> > +	mutex_unlock(&i915->perf.lock);
> >   	return ret;
> >   }
> > @@ -2511,11 +2511,11 @@ static void i915_perf_destroy_locked(struct i915_perf_stream *stream)
> >   static int i915_perf_release(struct inode *inode, struct file *file)
> >   {
> >   	struct i915_perf_stream *stream = file->private_data;
> > -	struct drm_i915_private *dev_priv = stream->dev_priv;
> > +	struct drm_i915_private *i915 = stream->i915;
> > -	mutex_lock(&dev_priv->perf.lock);
> > +	mutex_lock(&i915->perf.lock);
> >   	i915_perf_destroy_locked(stream);
> > -	mutex_unlock(&dev_priv->perf.lock);
> > +	mutex_unlock(&i915->perf.lock);
> >   	return 0;
> >   }
> > @@ -2537,7 +2537,7 @@ static const struct file_operations fops = {
> >   /**
> >    * i915_perf_open_ioctl_locked - DRM ioctl() for userspace to open a stream FD
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @param: The open parameters passed to 'DRM_I915_PERF_OPEN`
> >    * @props: individually validated u64 property value pairs
> >    * @file: drm file
> > @@ -2560,7 +2560,7 @@ static const struct file_operations fops = {
> >    * Returns: zero on success or a negative error code.
> >    */
> >   static int
> > -i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> > +i915_perf_open_ioctl_locked(struct drm_i915_private *i915,
> >   			    struct drm_i915_perf_open_param *param,
> >   			    struct perf_open_properties *props,
> >   			    struct drm_file *file)
> > @@ -2599,7 +2599,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> >   	 * MI_REPORT_PERF_COUNT commands and so consider it a privileged op to
> >   	 * enable the OA unit by default.
> >   	 */
> > -	if (IS_HASWELL(dev_priv) && specific_ctx)
> > +	if (IS_HASWELL(i915) && specific_ctx)
> >   		privileged_op = false;
> >   	/* Similar to perf's kernel.perf_paranoid_cpu sysctl option
> > @@ -2620,7 +2620,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> >   		goto err_ctx;
> >   	}
> > -	stream->dev_priv = dev_priv;
> > +	stream->i915 = i915;
> >   	stream->ctx = specific_ctx;
> >   	ret = i915_oa_stream_init(stream, param, props);
> > @@ -2636,7 +2636,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> >   		goto err_flags;
> >   	}
> > -	list_add(&stream->link, &dev_priv->perf.streams);
> > +	list_add(&stream->link, &i915->perf.streams);
> >   	if (param->flags & I915_PERF_FLAG_FD_CLOEXEC)
> >   		f_flags |= O_CLOEXEC;
> > @@ -2668,15 +2668,15 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv,
> >   	return ret;
> >   }
> > -static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent)
> > +static u64 oa_exponent_to_ns(struct drm_i915_private *i915, int exponent)
> >   {
> >   	return div64_u64(1000000000ULL * (2ULL << exponent),
> > -			 1000ULL * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz);
> > +			 1000ULL * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz);
> >   }
> >   /**
> >    * read_properties_unlocked - validate + copy userspace stream open properties
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @uprops: The array of u64 key value pairs given by userspace
> >    * @n_props: The number of key value pairs expected in @uprops
> >    * @props: The stream configuration built up while validating properties
> > @@ -2689,7 +2689,7 @@ static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent)
> >    * we shouldn't validate or assume anything about ordering here. This doesn't
> >    * rule out defining new properties with ordering requirements in the future.
> >    */
> > -static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> > +static int read_properties_unlocked(struct drm_i915_private *i915,
> >   				    u64 __user *uprops,
> >   				    u32 n_props,
> >   				    struct perf_open_properties *props)
> > @@ -2755,7 +2755,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> >   					  value);
> >   				return -EINVAL;
> >   			}
> > -			if (!dev_priv->perf.oa.oa_formats[value].size) {
> > +			if (!i915->perf.oa.oa_formats[value].size) {
> >   				DRM_DEBUG("Unsupported OA report format %llu\n",
> >   					  value);
> >   				return -EINVAL;
> > @@ -2776,7 +2776,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> >   			 */
> >   			BUILD_BUG_ON(sizeof(oa_period) != 8);
> > -			oa_period = oa_exponent_to_ns(dev_priv, value);
> > +			oa_period = oa_exponent_to_ns(i915, value);
> >   			/* This check is primarily to ensure that oa_period <=
> >   			 * UINT32_MAX (before passing to do_div which only
> > @@ -2839,13 +2839,13 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv,
> >   int i915_perf_open_ioctl(struct drm_device *dev, void *data,
> >   			 struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = dev->dev_private;
> > +	struct drm_i915_private *i915 = dev->dev_private;
> >   	struct drm_i915_perf_open_param *param = data;
> >   	struct perf_open_properties props;
> >   	u32 known_open_flags;
> >   	int ret;
> > -	if (!dev_priv->perf.initialized) {
> > +	if (!i915->perf.initialized) {
> >   		DRM_DEBUG("i915 perf interface not available for this system\n");
> >   		return -ENOTSUPP;
> >   	}
> > @@ -2858,124 +2858,124 @@ int i915_perf_open_ioctl(struct drm_device *dev, void *data,
> >   		return -EINVAL;
> >   	}
> > -	ret = read_properties_unlocked(dev_priv,
> > +	ret = read_properties_unlocked(i915,
> >   				       u64_to_user_ptr(param->properties_ptr),
> >   				       param->num_properties,
> >   				       &props);
> >   	if (ret)
> >   		return ret;
> > -	mutex_lock(&dev_priv->perf.lock);
> > -	ret = i915_perf_open_ioctl_locked(dev_priv, param, &props, file);
> > -	mutex_unlock(&dev_priv->perf.lock);
> > +	mutex_lock(&i915->perf.lock);
> > +	ret = i915_perf_open_ioctl_locked(i915, param, &props, file);
> > +	mutex_unlock(&i915->perf.lock);
> >   	return ret;
> >   }
> >   /**
> >    * i915_perf_register - exposes i915-perf to userspace
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * In particular OA metric sets are advertised under a sysfs metrics/
> >    * directory allowing userspace to enumerate valid IDs that can be
> >    * used to open an i915-perf stream.
> >    */
> > -void i915_perf_register(struct drm_i915_private *dev_priv)
> > +void i915_perf_register(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	if (!dev_priv->perf.initialized)
> > +	if (!i915->perf.initialized)
> >   		return;
> >   	/* To be sure we're synchronized with an attempted
> >   	 * i915_perf_open_ioctl(); considering that we register after
> >   	 * being exposed to userspace.
> >   	 */
> > -	mutex_lock(&dev_priv->perf.lock);
> > +	mutex_lock(&i915->perf.lock);
> > -	dev_priv->perf.metrics_kobj =
> > +	i915->perf.metrics_kobj =
> >   		kobject_create_and_add("metrics",
> > -				       &dev_priv->drm.primary->kdev->kobj);
> > -	if (!dev_priv->perf.metrics_kobj)
> > +				       &i915->drm.primary->kdev->kobj);
> > +	if (!i915->perf.metrics_kobj)
> >   		goto exit;
> > -	sysfs_attr_init(&dev_priv->perf.oa.test_config.sysfs_metric_id.attr);
> > -
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		i915_perf_load_test_config_icl(dev_priv);
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > -		i915_perf_load_test_config_cnl(dev_priv);
> > -	} else if (IS_COFFEELAKE(dev_priv)) {
> > -		if (IS_CFL_GT2(dev_priv))
> > -			i915_perf_load_test_config_cflgt2(dev_priv);
> > -		if (IS_CFL_GT3(dev_priv))
> > -			i915_perf_load_test_config_cflgt3(dev_priv);
> > -	} else if (IS_GEMINILAKE(dev_priv)) {
> > -		i915_perf_load_test_config_glk(dev_priv);
> > -	} else if (IS_KABYLAKE(dev_priv)) {
> > -		if (IS_KBL_GT2(dev_priv))
> > -			i915_perf_load_test_config_kblgt2(dev_priv);
> > -		else if (IS_KBL_GT3(dev_priv))
> > -			i915_perf_load_test_config_kblgt3(dev_priv);
> > -	} else if (IS_BROXTON(dev_priv)) {
> > -		i915_perf_load_test_config_bxt(dev_priv);
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > -		if (IS_SKL_GT2(dev_priv))
> > -			i915_perf_load_test_config_sklgt2(dev_priv);
> > -		else if (IS_SKL_GT3(dev_priv))
> > -			i915_perf_load_test_config_sklgt3(dev_priv);
> > -		else if (IS_SKL_GT4(dev_priv))
> > -			i915_perf_load_test_config_sklgt4(dev_priv);
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > -		i915_perf_load_test_config_chv(dev_priv);
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		i915_perf_load_test_config_bdw(dev_priv);
> > -	} else if (IS_HASWELL(dev_priv)) {
> > -		i915_perf_load_test_config_hsw(dev_priv);
> > -}
> > -
> > -	if (dev_priv->perf.oa.test_config.id == 0)
> > +	sysfs_attr_init(&i915->perf.oa.test_config.sysfs_metric_id.attr);
> > +
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		i915_perf_load_test_config_icl(i915);
> > +	} else if (IS_CANNONLAKE(i915)) {
> > +		i915_perf_load_test_config_cnl(i915);
> > +	} else if (IS_COFFEELAKE(i915)) {
> > +		if (IS_CFL_GT2(i915))
> > +			i915_perf_load_test_config_cflgt2(i915);
> > +		if (IS_CFL_GT3(i915))
> > +			i915_perf_load_test_config_cflgt3(i915);
> > +	} else if (IS_GEMINILAKE(i915)) {
> > +		i915_perf_load_test_config_glk(i915);
> > +	} else if (IS_KABYLAKE(i915)) {
> > +		if (IS_KBL_GT2(i915))
> > +			i915_perf_load_test_config_kblgt2(i915);
> > +		else if (IS_KBL_GT3(i915))
> > +			i915_perf_load_test_config_kblgt3(i915);
> > +	} else if (IS_BROXTON(i915)) {
> > +		i915_perf_load_test_config_bxt(i915);
> > +	} else if (IS_SKYLAKE(i915)) {
> > +		if (IS_SKL_GT2(i915))
> > +			i915_perf_load_test_config_sklgt2(i915);
> > +		else if (IS_SKL_GT3(i915))
> > +			i915_perf_load_test_config_sklgt3(i915);
> > +		else if (IS_SKL_GT4(i915))
> > +			i915_perf_load_test_config_sklgt4(i915);
> > +	} else if (IS_CHERRYVIEW(i915)) {
> > +		i915_perf_load_test_config_chv(i915);
> > +	} else if (IS_BROADWELL(i915)) {
> > +		i915_perf_load_test_config_bdw(i915);
> > +	} else if (IS_HASWELL(i915)) {
> > +		i915_perf_load_test_config_hsw(i915);
> > +}
> > +
> > +	if (i915->perf.oa.test_config.id == 0)
> >   		goto sysfs_error;
> > -	ret = sysfs_create_group(dev_priv->perf.metrics_kobj,
> > -				 &dev_priv->perf.oa.test_config.sysfs_metric);
> > +	ret = sysfs_create_group(i915->perf.metrics_kobj,
> > +				 &i915->perf.oa.test_config.sysfs_metric);
> >   	if (ret)
> >   		goto sysfs_error;
> > -	atomic_set(&dev_priv->perf.oa.test_config.ref_count, 1);
> > +	atomic_set(&i915->perf.oa.test_config.ref_count, 1);
> >   	goto exit;
> >   sysfs_error:
> > -	kobject_put(dev_priv->perf.metrics_kobj);
> > -	dev_priv->perf.metrics_kobj = NULL;
> > +	kobject_put(i915->perf.metrics_kobj);
> > +	i915->perf.metrics_kobj = NULL;
> >   exit:
> > -	mutex_unlock(&dev_priv->perf.lock);
> > +	mutex_unlock(&i915->perf.lock);
> >   }
> >   /**
> >    * i915_perf_unregister - hide i915-perf from userspace
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * i915-perf state cleanup is split up into an 'unregister' and
> >    * 'deinit' phase where the interface is first hidden from
> >    * userspace by i915_perf_unregister() before cleaning up
> >    * remaining state in i915_perf_fini().
> >    */
> > -void i915_perf_unregister(struct drm_i915_private *dev_priv)
> > +void i915_perf_unregister(struct drm_i915_private *i915)
> >   {
> > -	if (!dev_priv->perf.metrics_kobj)
> > +	if (!i915->perf.metrics_kobj)
> >   		return;
> > -	sysfs_remove_group(dev_priv->perf.metrics_kobj,
> > -			   &dev_priv->perf.oa.test_config.sysfs_metric);
> > +	sysfs_remove_group(i915->perf.metrics_kobj,
> > +			   &i915->perf.oa.test_config.sysfs_metric);
> > -	kobject_put(dev_priv->perf.metrics_kobj);
> > -	dev_priv->perf.metrics_kobj = NULL;
> > +	kobject_put(i915->perf.metrics_kobj);
> > +	i915->perf.metrics_kobj = NULL;
> >   }
> > -static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool gen8_is_valid_flex_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> >   	static const i915_reg_t flex_eu_regs[] = {
> >   		EU_PERF_CNTL0,
> > @@ -2995,7 +2995,7 @@ static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr)
> >   	return false;
> >   }
> > -static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> >   	return (addr >= i915_mmio_reg_offset(OASTARTTRIG1) &&
> >   		addr <= i915_mmio_reg_offset(OASTARTTRIG8)) ||
> > @@ -3005,7 +3005,7 @@ static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32
> >   		 addr <= i915_mmio_reg_offset(OACEC7_1));
> >   }
> > -static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool gen7_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> >   	return addr == i915_mmio_reg_offset(HALF_SLICE_CHICKEN2) ||
> >   		(addr >= i915_mmio_reg_offset(MICRO_BP0_0) &&
> > @@ -3016,34 +3016,34 @@ static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> >   		 addr <= i915_mmio_reg_offset(OA_PERFMATRIX_HI));
> >   }
> > -static bool gen8_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool gen8_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> > -	return gen7_is_valid_mux_addr(dev_priv, addr) ||
> > +	return gen7_is_valid_mux_addr(i915, addr) ||
> >   		addr == i915_mmio_reg_offset(WAIT_FOR_RC6_EXIT) ||
> >   		(addr >= i915_mmio_reg_offset(RPM_CONFIG0) &&
> >   		 addr <= i915_mmio_reg_offset(NOA_CONFIG(8)));
> >   }
> > -static bool gen10_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool gen10_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> > -	return gen8_is_valid_mux_addr(dev_priv, addr) ||
> > +	return gen8_is_valid_mux_addr(i915, addr) ||
> >   		addr == i915_mmio_reg_offset(GEN10_NOA_WRITE_HIGH) ||
> >   		(addr >= i915_mmio_reg_offset(OA_PERFCNT3_LO) &&
> >   		 addr <= i915_mmio_reg_offset(OA_PERFCNT4_HI));
> >   }
> > -static bool hsw_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool hsw_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> > -	return gen7_is_valid_mux_addr(dev_priv, addr) ||
> > +	return gen7_is_valid_mux_addr(i915, addr) ||
> >   		(addr >= 0x25100 && addr <= 0x2FF90) ||
> >   		(addr >= i915_mmio_reg_offset(HSW_MBVID2_NOA0) &&
> >   		 addr <= i915_mmio_reg_offset(HSW_MBVID2_NOA9)) ||
> >   		addr == i915_mmio_reg_offset(HSW_MBVID2_MISR0);
> >   }
> > -static bool chv_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr)
> > +static bool chv_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr)
> >   {
> > -	return gen7_is_valid_mux_addr(dev_priv, addr) ||
> > +	return gen7_is_valid_mux_addr(i915, addr) ||
> >   		(addr >= 0x182300 && addr <= 0x1823A4);
> >   }
> > @@ -3066,8 +3066,8 @@ static u32 mask_reg_value(u32 reg, u32 val)
> >   	return val;
> >   }
> > -static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv,
> > -					 bool (*is_valid)(struct drm_i915_private *dev_priv, u32 addr),
> > +static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *i915,
> > +					 bool (*is_valid)(struct drm_i915_private *i915, u32 addr),
> >   					 u32 __user *regs,
> >   					 u32 n_regs)
> >   {
> > @@ -3097,7 +3097,7 @@ static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv,
> >   		if (err)
> >   			goto addr_err;
> > -		if (!is_valid(dev_priv, addr)) {
> > +		if (!is_valid(i915, addr)) {
> >   			DRM_DEBUG("Invalid oa_reg address: %X\n", addr);
> >   			err = -EINVAL;
> >   			goto addr_err;
> > @@ -3130,7 +3130,7 @@ static ssize_t show_dynamic_id(struct device *dev,
> >   	return sprintf(buf, "%d\n", oa_config->id);
> >   }
> > -static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> > +static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *i915,
> >   					 struct i915_oa_config *oa_config)
> >   {
> >   	sysfs_attr_init(&oa_config->sysfs_metric_id.attr);
> > @@ -3145,7 +3145,7 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> >   	oa_config->sysfs_metric.name = oa_config->uuid;
> >   	oa_config->sysfs_metric.attrs = oa_config->attrs;
> > -	return sysfs_create_group(dev_priv->perf.metrics_kobj,
> > +	return sysfs_create_group(i915->perf.metrics_kobj,
> >   				  &oa_config->sysfs_metric);
> >   }
> > @@ -3165,17 +3165,17 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv,
> >   int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   			       struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = dev->dev_private;
> > +	struct drm_i915_private *i915 = dev->dev_private;
> >   	struct drm_i915_perf_oa_config *args = data;
> >   	struct i915_oa_config *oa_config, *tmp;
> >   	int err, id;
> > -	if (!dev_priv->perf.initialized) {
> > +	if (!i915->perf.initialized) {
> >   		DRM_DEBUG("i915 perf interface not available for this system\n");
> >   		return -ENOTSUPP;
> >   	}
> > -	if (!dev_priv->perf.metrics_kobj) {
> > +	if (!i915->perf.metrics_kobj) {
> >   		DRM_DEBUG("OA metrics weren't advertised via sysfs\n");
> >   		return -EINVAL;
> >   	}
> > @@ -3213,8 +3213,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   	oa_config->mux_regs_len = args->n_mux_regs;
> >   	oa_config->mux_regs =
> > -		alloc_oa_regs(dev_priv,
> > -			      dev_priv->perf.oa.ops.is_valid_mux_reg,
> > +		alloc_oa_regs(i915,
> > +			      i915->perf.oa.ops.is_valid_mux_reg,
> >   			      u64_to_user_ptr(args->mux_regs_ptr),
> >   			      args->n_mux_regs);
> > @@ -3226,8 +3226,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   	oa_config->b_counter_regs_len = args->n_boolean_regs;
> >   	oa_config->b_counter_regs =
> > -		alloc_oa_regs(dev_priv,
> > -			      dev_priv->perf.oa.ops.is_valid_b_counter_reg,
> > +		alloc_oa_regs(i915,
> > +			      i915->perf.oa.ops.is_valid_b_counter_reg,
> >   			      u64_to_user_ptr(args->boolean_regs_ptr),
> >   			      args->n_boolean_regs);
> > @@ -3237,7 +3237,7 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   		goto reg_err;
> >   	}
> > -	if (INTEL_GEN(dev_priv) < 8) {
> > +	if (INTEL_GEN(i915) < 8) {
> >   		if (args->n_flex_regs != 0) {
> >   			err = -EINVAL;
> >   			goto reg_err;
> > @@ -3245,8 +3245,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   	} else {
> >   		oa_config->flex_regs_len = args->n_flex_regs;
> >   		oa_config->flex_regs =
> > -			alloc_oa_regs(dev_priv,
> > -				      dev_priv->perf.oa.ops.is_valid_flex_reg,
> > +			alloc_oa_regs(i915,
> > +				      i915->perf.oa.ops.is_valid_flex_reg,
> >   				      u64_to_user_ptr(args->flex_regs_ptr),
> >   				      args->n_flex_regs);
> > @@ -3257,14 +3257,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   		}
> >   	}
> > -	err = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> > +	err = mutex_lock_interruptible(&i915->perf.metrics_lock);
> >   	if (err)
> >   		goto reg_err;
> >   	/* We shouldn't have too many configs, so this iteration shouldn't be
> >   	 * too costly.
> >   	 */
> > -	idr_for_each_entry(&dev_priv->perf.metrics_idr, tmp, id) {
> > +	idr_for_each_entry(&i915->perf.metrics_idr, tmp, id) {
> >   		if (!strcmp(tmp->uuid, oa_config->uuid)) {
> >   			DRM_DEBUG("OA config already exists with this uuid\n");
> >   			err = -EADDRINUSE;
> > @@ -3272,14 +3272,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   		}
> >   	}
> > -	err = create_dynamic_oa_sysfs_entry(dev_priv, oa_config);
> > +	err = create_dynamic_oa_sysfs_entry(i915, oa_config);
> >   	if (err) {
> >   		DRM_DEBUG("Failed to create sysfs entry for OA config\n");
> >   		goto sysfs_err;
> >   	}
> >   	/* Config id 0 is invalid, id 1 for kernel stored test config. */
> > -	oa_config->id = idr_alloc(&dev_priv->perf.metrics_idr,
> > +	oa_config->id = idr_alloc(&i915->perf.metrics_idr,
> >   				  oa_config, 2,
> >   				  0, GFP_KERNEL);
> >   	if (oa_config->id < 0) {
> > @@ -3288,16 +3288,16 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   		goto sysfs_err;
> >   	}
> > -	mutex_unlock(&dev_priv->perf.metrics_lock);
> > +	mutex_unlock(&i915->perf.metrics_lock);
> >   	DRM_DEBUG("Added config %s id=%i\n", oa_config->uuid, oa_config->id);
> >   	return oa_config->id;
> >   sysfs_err:
> > -	mutex_unlock(&dev_priv->perf.metrics_lock);
> > +	mutex_unlock(&i915->perf.metrics_lock);
> >   reg_err:
> > -	put_oa_config(dev_priv, oa_config);
> > +	put_oa_config(i915, oa_config);
> >   	DRM_DEBUG("Failed to add new OA config\n");
> >   	return err;
> >   }
> > @@ -3316,12 +3316,12 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
> >   int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
> >   				  struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = dev->dev_private;
> > +	struct drm_i915_private *i915 = dev->dev_private;
> >   	u64 *arg = data;
> >   	struct i915_oa_config *oa_config;
> >   	int ret;
> > -	if (!dev_priv->perf.initialized) {
> > +	if (!i915->perf.initialized) {
> >   		DRM_DEBUG("i915 perf interface not available for this system\n");
> >   		return -ENOTSUPP;
> >   	}
> > @@ -3331,11 +3331,11 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
> >   		return -EACCES;
> >   	}
> > -	ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock);
> > +	ret = mutex_lock_interruptible(&i915->perf.metrics_lock);
> >   	if (ret)
> >   		goto lock_err;
> > -	oa_config = idr_find(&dev_priv->perf.metrics_idr, *arg);
> > +	oa_config = idr_find(&i915->perf.metrics_idr, *arg);
> >   	if (!oa_config) {
> >   		DRM_DEBUG("Failed to remove unknown OA config\n");
> >   		ret = -ENOENT;
> > @@ -3344,17 +3344,17 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data,
> >   	GEM_BUG_ON(*arg != oa_config->id);
> > -	sysfs_remove_group(dev_priv->perf.metrics_kobj,
> > +	sysfs_remove_group(i915->perf.metrics_kobj,
> >   			   &oa_config->sysfs_metric);
> > -	idr_remove(&dev_priv->perf.metrics_idr, *arg);
> > +	idr_remove(&i915->perf.metrics_idr, *arg);
> >   	DRM_DEBUG("Removed config %s id=%i\n", oa_config->uuid, oa_config->id);
> > -	put_oa_config(dev_priv, oa_config);
> > +	put_oa_config(i915, oa_config);
> >   config_err:
> > -	mutex_unlock(&dev_priv->perf.metrics_lock);
> > +	mutex_unlock(&i915->perf.metrics_lock);
> >   lock_err:
> >   	return ret;
> >   }
> > @@ -3403,135 +3403,135 @@ static struct ctl_table dev_root[] = {
> >   /**
> >    * i915_perf_init - initialize i915-perf state on module load
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Initializes i915-perf state without exposing anything to userspace.
> >    *
> >    * Note: i915-perf initialization is split into an 'init' and 'register'
> >    * phase with the i915_perf_register() exposing state to userspace.
> >    */
> > -void i915_perf_init(struct drm_i915_private *dev_priv)
> > +void i915_perf_init(struct drm_i915_private *i915)
> >   {
> > -	if (IS_HASWELL(dev_priv)) {
> > -		dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> > +	if (IS_HASWELL(i915)) {
> > +		i915->perf.oa.ops.is_valid_b_counter_reg =
> >   			gen7_is_valid_b_counter_addr;
> > -		dev_priv->perf.oa.ops.is_valid_mux_reg =
> > +		i915->perf.oa.ops.is_valid_mux_reg =
> >   			hsw_is_valid_mux_addr;
> > -		dev_priv->perf.oa.ops.is_valid_flex_reg = NULL;
> > -		dev_priv->perf.oa.ops.enable_metric_set = hsw_enable_metric_set;
> > -		dev_priv->perf.oa.ops.disable_metric_set = hsw_disable_metric_set;
> > -		dev_priv->perf.oa.ops.oa_enable = gen7_oa_enable;
> > -		dev_priv->perf.oa.ops.oa_disable = gen7_oa_disable;
> > -		dev_priv->perf.oa.ops.read = gen7_oa_read;
> > -		dev_priv->perf.oa.ops.oa_hw_tail_read =
> > +		i915->perf.oa.ops.is_valid_flex_reg = NULL;
> > +		i915->perf.oa.ops.enable_metric_set = hsw_enable_metric_set;
> > +		i915->perf.oa.ops.disable_metric_set = hsw_disable_metric_set;
> > +		i915->perf.oa.ops.oa_enable = gen7_oa_enable;
> > +		i915->perf.oa.ops.oa_disable = gen7_oa_disable;
> > +		i915->perf.oa.ops.read = gen7_oa_read;
> > +		i915->perf.oa.ops.oa_hw_tail_read =
> >   			gen7_oa_hw_tail_read;
> > -		dev_priv->perf.oa.oa_formats = hsw_oa_formats;
> > -	} else if (HAS_LOGICAL_RING_CONTEXTS(dev_priv)) {
> > +		i915->perf.oa.oa_formats = hsw_oa_formats;
> > +	} else if (HAS_LOGICAL_RING_CONTEXTS(i915)) {
> >   		/* Note: that although we could theoretically also support the
> >   		 * legacy ringbuffer mode on BDW (and earlier iterations of
> >   		 * this driver, before upstreaming did this) it didn't seem
> >   		 * worth the complexity to maintain now that BDW+ enable
> >   		 * execlist mode by default.
> >   		 */
> > -		dev_priv->perf.oa.oa_formats = gen8_plus_oa_formats;
> > +		i915->perf.oa.oa_formats = gen8_plus_oa_formats;
> > -		dev_priv->perf.oa.ops.oa_enable = gen8_oa_enable;
> > -		dev_priv->perf.oa.ops.oa_disable = gen8_oa_disable;
> > -		dev_priv->perf.oa.ops.read = gen8_oa_read;
> > -		dev_priv->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read;
> > +		i915->perf.oa.ops.oa_enable = gen8_oa_enable;
> > +		i915->perf.oa.ops.oa_disable = gen8_oa_disable;
> > +		i915->perf.oa.ops.read = gen8_oa_read;
> > +		i915->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read;
> > -		if (IS_GEN_RANGE(dev_priv, 8, 9)) {
> > -			dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> > +		if (IS_GEN_RANGE(i915, 8, 9)) {
> > +			i915->perf.oa.ops.is_valid_b_counter_reg =
> >   				gen7_is_valid_b_counter_addr;
> > -			dev_priv->perf.oa.ops.is_valid_mux_reg =
> > +			i915->perf.oa.ops.is_valid_mux_reg =
> >   				gen8_is_valid_mux_addr;
> > -			dev_priv->perf.oa.ops.is_valid_flex_reg =
> > +			i915->perf.oa.ops.is_valid_flex_reg =
> >   				gen8_is_valid_flex_addr;
> > -			if (IS_CHERRYVIEW(dev_priv)) {
> > -				dev_priv->perf.oa.ops.is_valid_mux_reg =
> > +			if (IS_CHERRYVIEW(i915)) {
> > +				i915->perf.oa.ops.is_valid_mux_reg =
> >   					chv_is_valid_mux_addr;
> >   			}
> > -			dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> > -			dev_priv->perf.oa.ops.disable_metric_set = gen8_disable_metric_set;
> > +			i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> > +			i915->perf.oa.ops.disable_metric_set = gen8_disable_metric_set;
> > -			if (IS_GEN(dev_priv, 8)) {
> > -				dev_priv->perf.oa.ctx_oactxctrl_offset = 0x120;
> > -				dev_priv->perf.oa.ctx_flexeu0_offset = 0x2ce;
> > +			if (IS_GEN(i915, 8)) {
> > +				i915->perf.oa.ctx_oactxctrl_offset = 0x120;
> > +				i915->perf.oa.ctx_flexeu0_offset = 0x2ce;
> > -				dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<25);
> > +				i915->perf.oa.gen8_valid_ctx_bit = (1<<25);
> >   			} else {
> > -				dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128;
> > -				dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de;
> > +				i915->perf.oa.ctx_oactxctrl_offset = 0x128;
> > +				i915->perf.oa.ctx_flexeu0_offset = 0x3de;
> > -				dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16);
> > +				i915->perf.oa.gen8_valid_ctx_bit = (1<<16);
> >   			}
> > -		} else if (IS_GEN_RANGE(dev_priv, 10, 11)) {
> > -			dev_priv->perf.oa.ops.is_valid_b_counter_reg =
> > +		} else if (IS_GEN_RANGE(i915, 10, 11)) {
> > +			i915->perf.oa.ops.is_valid_b_counter_reg =
> >   				gen7_is_valid_b_counter_addr;
> > -			dev_priv->perf.oa.ops.is_valid_mux_reg =
> > +			i915->perf.oa.ops.is_valid_mux_reg =
> >   				gen10_is_valid_mux_addr;
> > -			dev_priv->perf.oa.ops.is_valid_flex_reg =
> > +			i915->perf.oa.ops.is_valid_flex_reg =
> >   				gen8_is_valid_flex_addr;
> > -			dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> > -			dev_priv->perf.oa.ops.disable_metric_set = gen10_disable_metric_set;
> > +			i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set;
> > +			i915->perf.oa.ops.disable_metric_set = gen10_disable_metric_set;
> > -			dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128;
> > -			dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de;
> > +			i915->perf.oa.ctx_oactxctrl_offset = 0x128;
> > +			i915->perf.oa.ctx_flexeu0_offset = 0x3de;
> > -			dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16);
> > +			i915->perf.oa.gen8_valid_ctx_bit = (1<<16);
> >   		}
> >   	}
> > -	if (dev_priv->perf.oa.ops.enable_metric_set) {
> > -		hrtimer_init(&dev_priv->perf.oa.poll_check_timer,
> > +	if (i915->perf.oa.ops.enable_metric_set) {
> > +		hrtimer_init(&i915->perf.oa.poll_check_timer,
> >   				CLOCK_MONOTONIC, HRTIMER_MODE_REL);
> > -		dev_priv->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb;
> > -		init_waitqueue_head(&dev_priv->perf.oa.poll_wq);
> > +		i915->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb;
> > +		init_waitqueue_head(&i915->perf.oa.poll_wq);
> > -		INIT_LIST_HEAD(&dev_priv->perf.streams);
> > -		mutex_init(&dev_priv->perf.lock);
> > -		spin_lock_init(&dev_priv->perf.oa.oa_buffer.ptr_lock);
> > +		INIT_LIST_HEAD(&i915->perf.streams);
> > +		mutex_init(&i915->perf.lock);
> > +		spin_lock_init(&i915->perf.oa.oa_buffer.ptr_lock);
> >   		oa_sample_rate_hard_limit = 1000 *
> > -			(RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz / 2);
> > -		dev_priv->perf.sysctl_header = register_sysctl_table(dev_root);
> > +			(RUNTIME_INFO(i915)->cs_timestamp_frequency_khz / 2);
> > +		i915->perf.sysctl_header = register_sysctl_table(dev_root);
> > -		mutex_init(&dev_priv->perf.metrics_lock);
> > -		idr_init(&dev_priv->perf.metrics_idr);
> > +		mutex_init(&i915->perf.metrics_lock);
> > +		idr_init(&i915->perf.metrics_idr);
> > -		dev_priv->perf.initialized = true;
> > +		i915->perf.initialized = true;
> >   	}
> >   }
> >   static int destroy_config(int id, void *p, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = data;
> > +	struct drm_i915_private *i915 = data;
> >   	struct i915_oa_config *oa_config = p;
> > -	put_oa_config(dev_priv, oa_config);
> > +	put_oa_config(i915, oa_config);
> >   	return 0;
> >   }
> >   /**
> >    * i915_perf_fini - Counter part to i915_perf_init()
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    */
> > -void i915_perf_fini(struct drm_i915_private *dev_priv)
> > +void i915_perf_fini(struct drm_i915_private *i915)
> >   {
> > -	if (!dev_priv->perf.initialized)
> > +	if (!i915->perf.initialized)
> >   		return;
> > -	idr_for_each(&dev_priv->perf.metrics_idr, destroy_config, dev_priv);
> > -	idr_destroy(&dev_priv->perf.metrics_idr);
> > +	idr_for_each(&i915->perf.metrics_idr, destroy_config, i915);
> > +	idr_destroy(&i915->perf.metrics_idr);
> > -	unregister_sysctl_table(dev_priv->perf.sysctl_header);
> > +	unregister_sysctl_table(i915->perf.sysctl_header);
> > -	memset(&dev_priv->perf.oa.ops, 0, sizeof(dev_priv->perf.oa.ops));
> > +	memset(&i915->perf.oa.ops, 0, sizeof(i915->perf.oa.ops));
> > -	dev_priv->perf.initialized = false;
> > +	i915->perf.initialized = false;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_pmu.c b/drivers/gpu/drm/i915/i915_pmu.c
> > index eb9c0e0e545c..c4861c12100c 100644
> > --- a/drivers/gpu/drm/i915/i915_pmu.c
> > +++ b/drivers/gpu/drm/i915/i915_pmu.c
> > @@ -159,24 +159,24 @@ add_sample(struct i915_pmu_sample *sample, u32 val)
> >   }
> >   static void
> > -engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> > +engines_sample(struct drm_i915_private *i915, unsigned int period_ns)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	intel_wakeref_t wakeref;
> >   	unsigned long flags;
> > -	if ((dev_priv->pmu.enable & ENGINE_SAMPLE_MASK) == 0)
> > +	if ((i915->pmu.enable & ENGINE_SAMPLE_MASK) == 0)
> >   		return;
> >   	wakeref = 0;
> > -	if (READ_ONCE(dev_priv->gt.awake))
> > -		wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> > +	if (READ_ONCE(i915->gt.awake))
> > +		wakeref = intel_runtime_pm_get_if_in_use(i915);
> >   	if (!wakeref)
> >   		return;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, flags);
> > -	for_each_engine(engine, dev_priv, id) {
> > +	spin_lock_irqsave(&i915->uncore.lock, flags);
> > +	for_each_engine(engine, i915, id) {
> >   		struct intel_engine_pmu *pmu = &engine->pmu;
> >   		bool busy;
> >   		u32 val;
> > @@ -205,9 +205,9 @@ engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> >   		if (busy)
> >   			add_sample(&pmu->sample[I915_SAMPLE_BUSY], period_ns);
> >   	}
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, flags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, flags);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   }
> >   static void
> > @@ -217,33 +217,33 @@ add_sample_mult(struct i915_pmu_sample *sample, u32 val, u32 mul)
> >   }
> >   static void
> > -frequency_sample(struct drm_i915_private *dev_priv, unsigned int period_ns)
> > +frequency_sample(struct drm_i915_private *i915, unsigned int period_ns)
> >   {
> > -	if (dev_priv->pmu.enable &
> > +	if (i915->pmu.enable &
> >   	    config_enabled_mask(I915_PMU_ACTUAL_FREQUENCY)) {
> >   		u32 val;
> > -		val = dev_priv->gt_pm.rps.cur_freq;
> > -		if (dev_priv->gt.awake) {
> > +		val = i915->gt_pm.rps.cur_freq;
> > +		if (i915->gt.awake) {
> >   			intel_wakeref_t wakeref;
> > -			with_intel_runtime_pm_if_in_use(dev_priv, wakeref) {
> > -				val = intel_uncore_read_notrace(&dev_priv->uncore,
> > +			with_intel_runtime_pm_if_in_use(i915, wakeref) {
> > +				val = intel_uncore_read_notrace(&i915->uncore,
> >   								GEN6_RPSTAT1);
> > -				val = intel_get_cagf(dev_priv, val);
> > +				val = intel_get_cagf(i915, val);
> >   			}
> >   		}
> > -		add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_ACT],
> > -				intel_gpu_freq(dev_priv, val),
> > +		add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_ACT],
> > +				intel_gpu_freq(i915, val),
> >   				period_ns / 1000);
> >   	}
> > -	if (dev_priv->pmu.enable &
> > +	if (i915->pmu.enable &
> >   	    config_enabled_mask(I915_PMU_REQUESTED_FREQUENCY)) {
> > -		add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_REQ],
> > -				intel_gpu_freq(dev_priv,
> > -					       dev_priv->gt_pm.rps.cur_freq),
> > +		add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_REQ],
> > +				intel_gpu_freq(i915,
> > +					       i915->gt_pm.rps.cur_freq),
> >   				period_ns / 1000);
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_query.c b/drivers/gpu/drm/i915/i915_query.c
> > index 7b7016171057..87606362b56f 100644
> > --- a/drivers/gpu/drm/i915/i915_query.c
> > +++ b/drivers/gpu/drm/i915/i915_query.c
> > @@ -31,10 +31,10 @@ static int copy_query_item(void *query_hdr, size_t query_sz,
> >   	return 0;
> >   }
> > -static int query_topology_info(struct drm_i915_private *dev_priv,
> > +static int query_topology_info(struct drm_i915_private *i915,
> >   			       struct drm_i915_query_item *query_item)
> >   {
> > -	const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	struct drm_i915_query_topology_info topo;
> >   	u32 slice_length, subslice_length, eu_length, total_length;
> >   	u8 subslice_stride = GEN_SSEU_STRIDE(sseu->max_subslices);
> > @@ -143,7 +143,7 @@ query_engine_info(struct drm_i915_private *i915,
> >   	return len;
> >   }
> > -static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv,
> > +static int (* const i915_query_funcs[])(struct drm_i915_private *i915,
> >   					struct drm_i915_query_item *query_item) = {
> >   	query_topology_info,
> >   	query_engine_info,
> > @@ -151,7 +151,7 @@ static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv,
> >   int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_query *args = data;
> >   	struct drm_i915_query_item __user *user_item_ptr =
> >   		u64_to_user_ptr(args->items_ptr);
> > @@ -180,7 +180,7 @@ int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
> >   		if (func_idx < ARRAY_SIZE(i915_query_funcs)) {
> >   			func_idx = array_index_nospec(func_idx,
> >   						      ARRAY_SIZE(i915_query_funcs));
> > -			ret = i915_query_funcs[func_idx](dev_priv, &item);
> > +			ret = i915_query_funcs[func_idx](i915, &item);
> >   		}
> >   		/* Only write the length back to userspace if they differ. */
> > diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h
> > index edf9f93934a1..b522970bdf74 100644
> > --- a/drivers/gpu/drm/i915/i915_reg.h
> > +++ b/drivers/gpu/drm/i915/i915_reg.h
> > @@ -205,7 +205,7 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg)
> >   #define VLV_MIPI_BASE			VLV_DISPLAY_BASE
> >   #define BXT_MIPI_BASE			0x60000
> > -#define DISPLAY_MMIO_BASE(dev_priv)	(INTEL_INFO(dev_priv)->display_mmio_offset)
> > +#define DISPLAY_MMIO_BASE(i915)	(INTEL_INFO(i915)->display_mmio_offset)
> >   /*
> >    * Given the first two numbers __a and __b of arbitrarily many evenly spaced
> > @@ -247,15 +247,15 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg)
> >    * Device info offset array based helpers for groups of registers with unevenly
> >    * spaced base offsets.
> >    */
> > -#define _MMIO_PIPE2(pipe, reg)		_MMIO(INTEL_INFO(dev_priv)->pipe_offsets[pipe] - \
> > -					      INTEL_INFO(dev_priv)->pipe_offsets[PIPE_A] + (reg) + \
> > -					      DISPLAY_MMIO_BASE(dev_priv))
> > -#define _MMIO_TRANS2(pipe, reg)		_MMIO(INTEL_INFO(dev_priv)->trans_offsets[(pipe)] - \
> > -					      INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_A] + (reg) + \
> > -					      DISPLAY_MMIO_BASE(dev_priv))
> > -#define _CURSOR2(pipe, reg)		_MMIO(INTEL_INFO(dev_priv)->cursor_offsets[(pipe)] - \
> > -					      INTEL_INFO(dev_priv)->cursor_offsets[PIPE_A] + (reg) + \
> > -					      DISPLAY_MMIO_BASE(dev_priv))
> > +#define _MMIO_PIPE2(pipe, reg)		_MMIO(INTEL_INFO(i915)->pipe_offsets[pipe] - \
> > +					      INTEL_INFO(i915)->pipe_offsets[PIPE_A] + (reg) + \
> > +					      DISPLAY_MMIO_BASE(i915))
> > +#define _MMIO_TRANS2(pipe, reg)		_MMIO(INTEL_INFO(i915)->trans_offsets[(pipe)] - \
> > +					      INTEL_INFO(i915)->trans_offsets[TRANSCODER_A] + (reg) + \
> > +					      DISPLAY_MMIO_BASE(i915))
> > +#define _CURSOR2(pipe, reg)		_MMIO(INTEL_INFO(i915)->cursor_offsets[(pipe)] - \
> > +					      INTEL_INFO(i915)->cursor_offsets[PIPE_A] + (reg) + \
> > +					      DISPLAY_MMIO_BASE(i915))
> >   #define __MASKED_FIELD(mask, value) ((mask) << 16 | (value))
> >   #define _MASKED_FIELD(mask, value) ({					   \
> > @@ -1307,7 +1307,7 @@ enum i915_power_well_id {
> >   #define  DPIO_CMNRST			(1 << 0)
> >   #define DPIO_PHY(pipe)			((pipe) >> 1)
> > -#define DPIO_PHY_IOSF_PORT(phy)		(dev_priv->dpio_phy_iosf_port[phy])
> > +#define DPIO_PHY_IOSF_PORT(phy)		(i915->dpio_phy_iosf_port[phy])
> >   /*
> >    * Per pipe/PLL DPIO regs
> > @@ -3004,9 +3004,9 @@ enum i915_power_well_id {
> >   #define PM_VEBOX_CS_ERROR_INTERRUPT		(1 << 12) /* hsw+ */
> >   #define PM_VEBOX_USER_INTERRUPT			(1 << 10) /* hsw+ */
> > -#define GT_PARITY_ERROR(dev_priv) \
> > +#define GT_PARITY_ERROR(i915) \
> >   	(GT_RENDER_L3_PARITY_ERROR_INTERRUPT | \
> > -	 (IS_HASWELL(dev_priv) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0))
> > +	 (IS_HASWELL(i915) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0))
> >   /* These are all the "old" interrupts */
> >   #define ILK_BSD_USER_INTERRUPT				(1 << 5)
> > @@ -3195,7 +3195,7 @@ enum i915_power_well_id {
> >   /*
> >    * GPIO regs
> >    */
> > -#define GPIO(gpio)		_MMIO(dev_priv->gpio_mmio_base + 0x5010 + \
> > +#define GPIO(gpio)		_MMIO(i915->gpio_mmio_base + 0x5010 + \
> >   				      4 * (gpio))
> >   # define GPIO_CLOCK_DIR_MASK		(1 << 0)
> > @@ -3213,7 +3213,7 @@ enum i915_power_well_id {
> >   # define GPIO_DATA_VAL_IN		(1 << 12)
> >   # define GPIO_DATA_PULLUP_DISABLE	(1 << 13)
> > -#define GMBUS0			_MMIO(dev_priv->gpio_mmio_base + 0x5100) /* clock/port select */
> > +#define GMBUS0			_MMIO(i915->gpio_mmio_base + 0x5100) /* clock/port select */
> >   #define   GMBUS_AKSV_SELECT	(1 << 11)
> >   #define   GMBUS_RATE_100KHZ	(0 << 8)
> >   #define   GMBUS_RATE_50KHZ	(1 << 8)
> > @@ -3240,7 +3240,7 @@ enum i915_power_well_id {
> >   #define   GMBUS_PIN_12_TC4_ICP	12
> >   #define   GMBUS_NUM_PINS	13 /* including 0 */
> > -#define GMBUS1			_MMIO(dev_priv->gpio_mmio_base + 0x5104) /* command/status */
> > +#define GMBUS1			_MMIO(i915->gpio_mmio_base + 0x5104) /* command/status */
> >   #define   GMBUS_SW_CLR_INT	(1 << 31)
> >   #define   GMBUS_SW_RDY		(1 << 30)
> >   #define   GMBUS_ENT		(1 << 29) /* enable timeout */
> > @@ -3255,7 +3255,7 @@ enum i915_power_well_id {
> >   #define   GMBUS_SLAVE_ADDR_SHIFT 1
> >   #define   GMBUS_SLAVE_READ	(1 << 0)
> >   #define   GMBUS_SLAVE_WRITE	(0 << 0)
> > -#define GMBUS2			_MMIO(dev_priv->gpio_mmio_base + 0x5108) /* status */
> > +#define GMBUS2			_MMIO(i915->gpio_mmio_base + 0x5108) /* status */
> >   #define   GMBUS_INUSE		(1 << 15)
> >   #define   GMBUS_HW_WAIT_PHASE	(1 << 14)
> >   #define   GMBUS_STALL_TIMEOUT	(1 << 13)
> > @@ -3263,22 +3263,22 @@ enum i915_power_well_id {
> >   #define   GMBUS_HW_RDY		(1 << 11)
> >   #define   GMBUS_SATOER		(1 << 10)
> >   #define   GMBUS_ACTIVE		(1 << 9)
> > -#define GMBUS3			_MMIO(dev_priv->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */
> > -#define GMBUS4			_MMIO(dev_priv->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */
> > +#define GMBUS3			_MMIO(i915->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */
> > +#define GMBUS4			_MMIO(i915->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */
> >   #define   GMBUS_SLAVE_TIMEOUT_EN (1 << 4)
> >   #define   GMBUS_NAK_EN		(1 << 3)
> >   #define   GMBUS_IDLE_EN		(1 << 2)
> >   #define   GMBUS_HW_WAIT_EN	(1 << 1)
> >   #define   GMBUS_HW_RDY_EN	(1 << 0)
> > -#define GMBUS5			_MMIO(dev_priv->gpio_mmio_base + 0x5120) /* byte index */
> > +#define GMBUS5			_MMIO(i915->gpio_mmio_base + 0x5120) /* byte index */
> >   #define   GMBUS_2BYTE_INDEX_EN	(1 << 31)
> >   /*
> >    * Clock control & power management
> >    */
> > -#define _DPLL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x6014)
> > -#define _DPLL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x6018)
> > -#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(dev_priv) + 0x6030)
> > +#define _DPLL_A (DISPLAY_MMIO_BASE(i915) + 0x6014)
> > +#define _DPLL_B (DISPLAY_MMIO_BASE(i915) + 0x6018)
> > +#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(i915) + 0x6030)
> >   #define DPLL(pipe) _MMIO_PIPE3((pipe), _DPLL_A, _DPLL_B, _CHV_DPLL_C)
> >   #define VGA0	_MMIO(0x6000)
> > @@ -3375,9 +3375,9 @@ enum i915_power_well_id {
> >   #define   SDVO_MULTIPLIER_SHIFT_HIRES		4
> >   #define   SDVO_MULTIPLIER_SHIFT_VGA		0
> > -#define _DPLL_A_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x601c)
> > -#define _DPLL_B_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x6020)
> > -#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x603c)
> > +#define _DPLL_A_MD (DISPLAY_MMIO_BASE(i915) + 0x601c)
> > +#define _DPLL_B_MD (DISPLAY_MMIO_BASE(i915) + 0x6020)
> > +#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(i915) + 0x603c)
> >   #define DPLL_MD(pipe) _MMIO_PIPE3((pipe), _DPLL_A_MD, _DPLL_B_MD, _CHV_DPLL_C_MD)
> >   /*
> > @@ -3449,7 +3449,7 @@ enum i915_power_well_id {
> >   #define  DSTATE_PLL_D3_OFF			(1 << 3)
> >   #define  DSTATE_GFX_CLOCK_GATING		(1 << 1)
> >   #define  DSTATE_DOT_CLOCK_GATING		(1 << 0)
> > -#define DSPCLK_GATE_D	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x6200)
> > +#define DSPCLK_GATE_D	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x6200)
> >   # define DPUNIT_B_CLOCK_GATE_DISABLE		(1 << 30) /* 965 */
> >   # define VSUNIT_CLOCK_GATE_DISABLE		(1 << 29) /* 965 */
> >   # define VRHUNIT_CLOCK_GATE_DISABLE		(1 << 28) /* 965 */
> > @@ -3589,7 +3589,7 @@ enum i915_power_well_id {
> >   #define _PALETTE_A		0xa000
> >   #define _PALETTE_B		0xa800
> >   #define _CHV_PALETTE_C		0xc000
> > -#define PALETTE(pipe, i)	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + \
> > +#define PALETTE(pipe, i)	_MMIO(DISPLAY_MMIO_BASE(i915) + \
> >   				      _PICK((pipe), _PALETTE_A,		\
> >   					    _PALETTE_B, _CHV_PALETTE_C) + \
> >   				      (i) * 4)
> > @@ -3936,8 +3936,8 @@ enum i915_power_well_id {
> >   #define INTERVAL_1_28_US(us)	roundup(((us) * 100) >> 7, 25)
> >   #define INTERVAL_1_33_US(us)	(((us) * 3)   >> 2)
> >   #define INTERVAL_0_833_US(us)	(((us) * 6) / 5)
> > -#define GT_INTERVAL_FROM_US(dev_priv, us) (INTEL_GEN(dev_priv) >= 9 ? \
> > -				(IS_GEN9_LP(dev_priv) ? \
> > +#define GT_INTERVAL_FROM_US(i915, us) (INTEL_GEN(i915) >= 9 ? \
> > +				(IS_GEN9_LP(i915) ? \
> >   				INTERVAL_0_833_US(us) : \
> >   				INTERVAL_1_33_US(us)) : \
> >   				INTERVAL_1_28_US(us))
> > @@ -3945,8 +3945,8 @@ enum i915_power_well_id {
> >   #define INTERVAL_1_28_TO_US(interval)  (((interval) << 7) / 100)
> >   #define INTERVAL_1_33_TO_US(interval)  (((interval) << 2) / 3)
> >   #define INTERVAL_0_833_TO_US(interval) (((interval) * 5)  / 6)
> > -#define GT_PM_INTERVAL_TO_US(dev_priv, interval) (INTEL_GEN(dev_priv) >= 9 ? \
> > -                           (IS_GEN9_LP(dev_priv) ? \
> > +#define GT_PM_INTERVAL_TO_US(i915, interval) (INTEL_GEN(i915) >= 9 ? \
> > +                           (IS_GEN9_LP(i915) ? \
> >                              INTERVAL_0_833_TO_US(interval) : \
> >                              INTERVAL_1_33_TO_US(interval)) : \
> >                              INTERVAL_1_28_TO_US(interval))
> > @@ -4219,7 +4219,7 @@ enum {
> >   /* HSW+ eDP PSR registers */
> >   #define HSW_EDP_PSR_BASE	0x64800
> >   #define BDW_EDP_PSR_BASE	0x6f800
> > -#define EDP_PSR_CTL				_MMIO(dev_priv->psr_mmio_base + 0)
> > +#define EDP_PSR_CTL				_MMIO(i915->psr_mmio_base + 0)
> >   #define   EDP_PSR_ENABLE			(1 << 31)
> >   #define   BDW_PSR_SINGLE_FRAME			(1 << 30)
> >   #define   EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK	(1 << 29) /* SW can't modify */
> > @@ -4256,16 +4256,16 @@ enum {
> >   #define   EDP_PSR_TRANSCODER_A_SHIFT		8
> >   #define   EDP_PSR_TRANSCODER_EDP_SHIFT		0
> > -#define EDP_PSR_AUX_CTL				_MMIO(dev_priv->psr_mmio_base + 0x10)
> > +#define EDP_PSR_AUX_CTL				_MMIO(i915->psr_mmio_base + 0x10)
> >   #define   EDP_PSR_AUX_CTL_TIME_OUT_MASK		(3 << 26)
> >   #define   EDP_PSR_AUX_CTL_MESSAGE_SIZE_MASK	(0x1f << 20)
> >   #define   EDP_PSR_AUX_CTL_PRECHARGE_2US_MASK	(0xf << 16)
> >   #define   EDP_PSR_AUX_CTL_ERROR_INTERRUPT	(1 << 11)
> >   #define   EDP_PSR_AUX_CTL_BIT_CLOCK_2X_MASK	(0x7ff)
> > -#define EDP_PSR_AUX_DATA(i)			_MMIO(dev_priv->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */
> > +#define EDP_PSR_AUX_DATA(i)			_MMIO(i915->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */
> > -#define EDP_PSR_STATUS				_MMIO(dev_priv->psr_mmio_base + 0x40)
> > +#define EDP_PSR_STATUS				_MMIO(i915->psr_mmio_base + 0x40)
> >   #define   EDP_PSR_STATUS_STATE_MASK		(7 << 29)
> >   #define   EDP_PSR_STATUS_STATE_SHIFT		29
> >   #define   EDP_PSR_STATUS_STATE_IDLE		(0 << 29)
> > @@ -4290,10 +4290,10 @@ enum {
> >   #define   EDP_PSR_STATUS_SENDING_TP1		(1 << 4)
> >   #define   EDP_PSR_STATUS_IDLE_MASK		0xf
> > -#define EDP_PSR_PERF_CNT		_MMIO(dev_priv->psr_mmio_base + 0x44)
> > +#define EDP_PSR_PERF_CNT		_MMIO(i915->psr_mmio_base + 0x44)
> >   #define   EDP_PSR_PERF_CNT_MASK		0xffffff
> > -#define EDP_PSR_DEBUG				_MMIO(dev_priv->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */
> > +#define EDP_PSR_DEBUG				_MMIO(i915->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */
> >   #define   EDP_PSR_DEBUG_MASK_MAX_SLEEP         (1 << 28)
> >   #define   EDP_PSR_DEBUG_MASK_LPSP              (1 << 27)
> >   #define   EDP_PSR_DEBUG_MASK_MEMUP             (1 << 26)
> > @@ -4405,7 +4405,7 @@ enum {
> >   /* Hotplug control (945+ only) */
> > -#define PORT_HOTPLUG_EN		_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61110)
> > +#define PORT_HOTPLUG_EN		_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61110)
> >   #define   PORTB_HOTPLUG_INT_EN			(1 << 29)
> >   #define   PORTC_HOTPLUG_INT_EN			(1 << 28)
> >   #define   PORTD_HOTPLUG_INT_EN			(1 << 27)
> > @@ -4435,7 +4435,7 @@ enum {
> >   #define CRT_HOTPLUG_DETECT_VOLTAGE_325MV	(0 << 2)
> >   #define CRT_HOTPLUG_DETECT_VOLTAGE_475MV	(1 << 2)
> > -#define PORT_HOTPLUG_STAT	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61114)
> > +#define PORT_HOTPLUG_STAT	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61114)
> >   /*
> >    * HDMI/DP bits are g4x+
> >    *
> > @@ -4517,7 +4517,7 @@ enum {
> >   #define PORT_DFT_I9XX				_MMIO(0x61150)
> >   #define   DC_BALANCE_RESET			(1 << 25)
> > -#define PORT_DFT2_G4X		_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61154)
> > +#define PORT_DFT2_G4X		_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61154)
> >   #define   DC_BALANCE_RESET_VLV			(1 << 31)
> >   #define   PIPE_SCRAMBLE_RESET_MASK		((1 << 14) | (0x3 << 0))
> >   #define   PIPE_C_SCRAMBLE_RESET			(1 << 14) /* chv */
> > @@ -4719,7 +4719,7 @@ enum {
> >   #define VLV_PPS_BASE			(VLV_DISPLAY_BASE + PPS_BASE)
> >   #define PCH_PPS_BASE			0xC7200
> > -#define _MMIO_PPS(pps_idx, reg)		_MMIO(dev_priv->pps_mmio_base -	\
> > +#define _MMIO_PPS(pps_idx, reg)		_MMIO(i915->pps_mmio_base -	\
> >   					      PPS_BASE + (reg) +	\
> >   					      (pps_idx) * 0x100)
> > @@ -4792,7 +4792,7 @@ enum {
> >   #define  PANEL_POWER_CYCLE_DELAY_MASK	REG_GENMASK(4, 0)
> >   /* Panel fitting */
> > -#define PFIT_CONTROL	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61230)
> > +#define PFIT_CONTROL	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61230)
> >   #define   PFIT_ENABLE		(1 << 31)
> >   #define   PFIT_PIPE_MASK	(3 << 29)
> >   #define   PFIT_PIPE_SHIFT	29
> > @@ -4810,7 +4810,7 @@ enum {
> >   #define   PFIT_SCALING_PROGRAMMED (1 << 26)
> >   #define   PFIT_SCALING_PILLAR	(2 << 26)
> >   #define   PFIT_SCALING_LETTER	(3 << 26)
> > -#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61234)
> > +#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61234)
> >   /* Pre-965 */
> >   #define		PFIT_VERT_SCALE_SHIFT		20
> >   #define		PFIT_VERT_SCALE_MASK		0xfff00000
> > @@ -4822,25 +4822,25 @@ enum {
> >   #define		PFIT_HORIZ_SCALE_SHIFT_965	0
> >   #define		PFIT_HORIZ_SCALE_MASK_965	0x00001fff
> > -#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61238)
> > +#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61238)
> > -#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61250)
> > -#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61350)
> > +#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(i915) + 0x61250)
> > +#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(i915) + 0x61350)
> >   #define VLV_BLC_PWM_CTL2(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL2_A, \
> >   					 _VLV_BLC_PWM_CTL2_B)
> > -#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61254)
> > -#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61354)
> > +#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61254)
> > +#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61354)
> >   #define VLV_BLC_PWM_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL_A, \
> >   					_VLV_BLC_PWM_CTL_B)
> > -#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61260)
> > -#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61360)
> > +#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61260)
> > +#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61360)
> >   #define VLV_BLC_HIST_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_HIST_CTL_A, \
> >   					 _VLV_BLC_HIST_CTL_B)
> >   /* Backlight control */
> > -#define BLC_PWM_CTL2	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61250) /* 965+ only */
> > +#define BLC_PWM_CTL2	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61250) /* 965+ only */
> >   #define   BLM_PWM_ENABLE		(1 << 31)
> >   #define   BLM_COMBINATION_MODE		(1 << 30) /* gen4 only */
> >   #define   BLM_PIPE_SELECT		(1 << 29)
> > @@ -4863,7 +4863,7 @@ enum {
> >   #define   BLM_PHASE_IN_COUNT_MASK	(0xff << 8)
> >   #define   BLM_PHASE_IN_INCR_SHIFT	(0)
> >   #define   BLM_PHASE_IN_INCR_MASK	(0xff << 0)
> > -#define BLC_PWM_CTL	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61254)
> > +#define BLC_PWM_CTL	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61254)
> >   /*
> >    * This is the most significant 15 bits of the number of backlight cycles in a
> >    * complete cycle of the modulated backlight control.
> > @@ -4885,7 +4885,7 @@ enum {
> >   #define   BACKLIGHT_DUTY_CYCLE_MASK_PNV		(0xfffe)
> >   #define   BLM_POLARITY_PNV			(1 << 0) /* pnv only */
> > -#define BLC_HIST_CTL	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61260)
> > +#define BLC_HIST_CTL	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x61260)
> >   #define  BLM_HISTOGRAM_ENABLE			(1 << 31)
> >   /* New registers for PCH-split platforms. Safe where new bits show up, the
> > @@ -5510,47 +5510,47 @@ enum {
> >    * is 20 bytes in each direction, hence the 5 fixed
> >    * data registers
> >    */
> > -#define _DPA_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64010)
> > -#define _DPA_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64014)
> > -#define _DPA_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64018)
> > -#define _DPA_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6401c)
> > -#define _DPA_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64020)
> > -#define _DPA_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64024)
> > -
> > -#define _DPB_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64110)
> > -#define _DPB_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64114)
> > -#define _DPB_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64118)
> > -#define _DPB_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6411c)
> > -#define _DPB_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64120)
> > -#define _DPB_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64124)
> > -
> > -#define _DPC_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64210)
> > -#define _DPC_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64214)
> > -#define _DPC_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64218)
> > -#define _DPC_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6421c)
> > -#define _DPC_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64220)
> > -#define _DPC_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64224)
> > -
> > -#define _DPD_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64310)
> > -#define _DPD_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64314)
> > -#define _DPD_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64318)
> > -#define _DPD_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6431c)
> > -#define _DPD_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64320)
> > -#define _DPD_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64324)
> > -
> > -#define _DPE_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64410)
> > -#define _DPE_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64414)
> > -#define _DPE_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64418)
> > -#define _DPE_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6441c)
> > -#define _DPE_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64420)
> > -#define _DPE_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64424)
> > -
> > -#define _DPF_AUX_CH_CTL		(DISPLAY_MMIO_BASE(dev_priv) + 0x64510)
> > -#define _DPF_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(dev_priv) + 0x64514)
> > -#define _DPF_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(dev_priv) + 0x64518)
> > -#define _DPF_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(dev_priv) + 0x6451c)
> > -#define _DPF_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(dev_priv) + 0x64520)
> > -#define _DPF_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(dev_priv) + 0x64524)
> > +#define _DPA_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64010)
> > +#define _DPA_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64014)
> > +#define _DPA_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64018)
> > +#define _DPA_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6401c)
> > +#define _DPA_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64020)
> > +#define _DPA_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64024)
> > +
> > +#define _DPB_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64110)
> > +#define _DPB_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64114)
> > +#define _DPB_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64118)
> > +#define _DPB_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6411c)
> > +#define _DPB_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64120)
> > +#define _DPB_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64124)
> > +
> > +#define _DPC_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64210)
> > +#define _DPC_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64214)
> > +#define _DPC_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64218)
> > +#define _DPC_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6421c)
> > +#define _DPC_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64220)
> > +#define _DPC_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64224)
> > +
> > +#define _DPD_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64310)
> > +#define _DPD_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64314)
> > +#define _DPD_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64318)
> > +#define _DPD_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6431c)
> > +#define _DPD_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64320)
> > +#define _DPD_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64324)
> > +
> > +#define _DPE_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64410)
> > +#define _DPE_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64414)
> > +#define _DPE_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64418)
> > +#define _DPE_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6441c)
> > +#define _DPE_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64420)
> > +#define _DPE_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64424)
> > +
> > +#define _DPF_AUX_CH_CTL		(DISPLAY_MMIO_BASE(i915) + 0x64510)
> > +#define _DPF_AUX_CH_DATA1	(DISPLAY_MMIO_BASE(i915) + 0x64514)
> > +#define _DPF_AUX_CH_DATA2	(DISPLAY_MMIO_BASE(i915) + 0x64518)
> > +#define _DPF_AUX_CH_DATA3	(DISPLAY_MMIO_BASE(i915) + 0x6451c)
> > +#define _DPF_AUX_CH_DATA4	(DISPLAY_MMIO_BASE(i915) + 0x64520)
> > +#define _DPF_AUX_CH_DATA5	(DISPLAY_MMIO_BASE(i915) + 0x64524)
> >   #define DP_AUX_CH_CTL(aux_ch)	_MMIO_PORT(aux_ch, _DPA_AUX_CH_CTL, _DPB_AUX_CH_CTL)
> >   #define DP_AUX_CH_DATA(aux_ch, i)	_MMIO(_PORT(aux_ch, _DPA_AUX_CH_DATA1, _DPB_AUX_CH_DATA1) + (i) * 4) /* 5 registers */
> > @@ -5843,7 +5843,7 @@ enum {
> >   #define   DPINVGTT_STATUS_MASK			0xff
> >   #define   DPINVGTT_STATUS_MASK_CHV		0xfff
> > -#define DSPARB			_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70030)
> > +#define DSPARB			_MMIO(DISPLAY_MMIO_BASE(i915) + 0x70030)
> >   #define   DSPARB_CSTART_MASK	(0x7f << 7)
> >   #define   DSPARB_CSTART_SHIFT	7
> >   #define   DSPARB_BSTART_MASK	(0x7f)
> > @@ -5878,7 +5878,7 @@ enum {
> >   #define   DSPARB_SPRITEF_MASK_VLV	(0xff << 8)
> >   /* pnv/gen4/g4x/vlv/chv */
> > -#define DSPFW1		_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70034)
> > +#define DSPFW1		_MMIO(DISPLAY_MMIO_BASE(i915) + 0x70034)
> >   #define   DSPFW_SR_SHIFT		23
> >   #define   DSPFW_SR_MASK			(0x1ff << 23)
> >   #define   DSPFW_CURSORB_SHIFT		16
> > @@ -5889,7 +5889,7 @@ enum {
> >   #define   DSPFW_PLANEA_SHIFT		0
> >   #define   DSPFW_PLANEA_MASK		(0x7f << 0)
> >   #define   DSPFW_PLANEA_MASK_VLV		(0xff << 0) /* vlv/chv */
> > -#define DSPFW2		_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70038)
> > +#define DSPFW2		_MMIO(DISPLAY_MMIO_BASE(i915) + 0x70038)
> >   #define   DSPFW_FBC_SR_EN		(1 << 31)	  /* g4x */
> >   #define   DSPFW_FBC_SR_SHIFT		28
> >   #define   DSPFW_FBC_SR_MASK		(0x7 << 28) /* g4x */
> > @@ -5905,7 +5905,7 @@ enum {
> >   #define   DSPFW_SPRITEA_SHIFT		0
> >   #define   DSPFW_SPRITEA_MASK		(0x7f << 0) /* g4x */
> >   #define   DSPFW_SPRITEA_MASK_VLV	(0xff << 0) /* vlv/chv */
> > -#define DSPFW3		_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x7003c)
> > +#define DSPFW3		_MMIO(DISPLAY_MMIO_BASE(i915) + 0x7003c)
> >   #define   DSPFW_HPLL_SR_EN		(1 << 31)
> >   #define   PINEVIEW_SELF_REFRESH_EN	(1 << 30)
> >   #define   DSPFW_CURSOR_SR_SHIFT		24
> > @@ -6322,35 +6322,35 @@ enum {
> >    * [10:1f] all
> >    * [30:32] all
> >    */
> > -#define SWF0(i)	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70410 + (i) * 4)
> > -#define SWF1(i)	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x71410 + (i) * 4)
> > -#define SWF3(i)	_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x72414 + (i) * 4)
> > +#define SWF0(i)	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x70410 + (i) * 4)
> > +#define SWF1(i)	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x71410 + (i) * 4)
> > +#define SWF3(i)	_MMIO(DISPLAY_MMIO_BASE(i915) + 0x72414 + (i) * 4)
> >   #define SWF_ILK(i)	_MMIO(0x4F000 + (i) * 4)
> >   /* Pipe B */
> > -#define _PIPEBDSL		(DISPLAY_MMIO_BASE(dev_priv) + 0x71000)
> > -#define _PIPEBCONF		(DISPLAY_MMIO_BASE(dev_priv) + 0x71008)
> > -#define _PIPEBSTAT		(DISPLAY_MMIO_BASE(dev_priv) + 0x71024)
> > +#define _PIPEBDSL		(DISPLAY_MMIO_BASE(i915) + 0x71000)
> > +#define _PIPEBCONF		(DISPLAY_MMIO_BASE(i915) + 0x71008)
> > +#define _PIPEBSTAT		(DISPLAY_MMIO_BASE(i915) + 0x71024)
> >   #define _PIPEBFRAMEHIGH		0x71040
> >   #define _PIPEBFRAMEPIXEL	0x71044
> > -#define _PIPEB_FRMCOUNT_G4X	(DISPLAY_MMIO_BASE(dev_priv) + 0x71040)
> > -#define _PIPEB_FLIPCOUNT_G4X	(DISPLAY_MMIO_BASE(dev_priv) + 0x71044)
> > +#define _PIPEB_FRMCOUNT_G4X	(DISPLAY_MMIO_BASE(i915) + 0x71040)
> > +#define _PIPEB_FLIPCOUNT_G4X	(DISPLAY_MMIO_BASE(i915) + 0x71044)
> >   /* Display B control */
> > -#define _DSPBCNTR		(DISPLAY_MMIO_BASE(dev_priv) + 0x71180)
> > +#define _DSPBCNTR		(DISPLAY_MMIO_BASE(i915) + 0x71180)
> >   #define   DISPPLANE_ALPHA_TRANS_ENABLE		(1 << 15)
> >   #define   DISPPLANE_ALPHA_TRANS_DISABLE		0
> >   #define   DISPPLANE_SPRITE_ABOVE_DISPLAY	0
> >   #define   DISPPLANE_SPRITE_ABOVE_OVERLAY	(1)
> > -#define _DSPBADDR		(DISPLAY_MMIO_BASE(dev_priv) + 0x71184)
> > -#define _DSPBSTRIDE		(DISPLAY_MMIO_BASE(dev_priv) + 0x71188)
> > -#define _DSPBPOS		(DISPLAY_MMIO_BASE(dev_priv) + 0x7118C)
> > -#define _DSPBSIZE		(DISPLAY_MMIO_BASE(dev_priv) + 0x71190)
> > -#define _DSPBSURF		(DISPLAY_MMIO_BASE(dev_priv) + 0x7119C)
> > -#define _DSPBTILEOFF		(DISPLAY_MMIO_BASE(dev_priv) + 0x711A4)
> > -#define _DSPBOFFSET		(DISPLAY_MMIO_BASE(dev_priv) + 0x711A4)
> > -#define _DSPBSURFLIVE		(DISPLAY_MMIO_BASE(dev_priv) + 0x711AC)
> > +#define _DSPBADDR		(DISPLAY_MMIO_BASE(i915) + 0x71184)
> > +#define _DSPBSTRIDE		(DISPLAY_MMIO_BASE(i915) + 0x71188)
> > +#define _DSPBPOS		(DISPLAY_MMIO_BASE(i915) + 0x7118C)
> > +#define _DSPBSIZE		(DISPLAY_MMIO_BASE(i915) + 0x71190)
> > +#define _DSPBSURF		(DISPLAY_MMIO_BASE(i915) + 0x7119C)
> > +#define _DSPBTILEOFF		(DISPLAY_MMIO_BASE(i915) + 0x711A4)
> > +#define _DSPBOFFSET		(DISPLAY_MMIO_BASE(i915) + 0x711A4)
> > +#define _DSPBSURFLIVE		(DISPLAY_MMIO_BASE(i915) + 0x711AC)
> >   /* ICL DSI 0 and 1 */
> >   #define _PIPEDSI0CONF		0x7b008
> > @@ -8946,7 +8946,7 @@ enum {
> >   #define   GEN9_ENABLE_GPGPU_PREEMPTION	(1 << 2)
> >   /* Audio */
> > -#define G4X_AUD_VID_DID			_MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x62020)
> > +#define G4X_AUD_VID_DID			_MMIO(DISPLAY_MMIO_BASE(i915) + 0x62020)
> >   #define   INTEL_AUDIO_DEVCL		0x808629FB
> >   #define   INTEL_AUDIO_DEVBLC		0x80862801
> >   #define   INTEL_AUDIO_DEVCTG		0x80862802
> > @@ -10469,8 +10469,8 @@ enum skl_power_gate {
> >   /* MIPI DSI Controller and D-PHY registers */
> > -#define _MIPIA_DEVICE_READY		(dev_priv->mipi_mmio_base + 0xb000)
> > -#define _MIPIC_DEVICE_READY		(dev_priv->mipi_mmio_base + 0xb800)
> > +#define _MIPIA_DEVICE_READY		(i915->mipi_mmio_base + 0xb000)
> > +#define _MIPIC_DEVICE_READY		(i915->mipi_mmio_base + 0xb800)
> >   #define MIPI_DEVICE_READY(port)		_MMIO_MIPI(port, _MIPIA_DEVICE_READY, _MIPIC_DEVICE_READY)
> >   #define  BUS_POSSESSION					(1 << 3) /* set to give bus to receiver */
> >   #define  ULPS_STATE_MASK				(3 << 1)
> > @@ -10479,11 +10479,11 @@ enum skl_power_gate {
> >   #define  ULPS_STATE_NORMAL_OPERATION			(0 << 1)
> >   #define  DEVICE_READY					(1 << 0)
> > -#define _MIPIA_INTR_STAT		(dev_priv->mipi_mmio_base + 0xb004)
> > -#define _MIPIC_INTR_STAT		(dev_priv->mipi_mmio_base + 0xb804)
> > +#define _MIPIA_INTR_STAT		(i915->mipi_mmio_base + 0xb004)
> > +#define _MIPIC_INTR_STAT		(i915->mipi_mmio_base + 0xb804)
> >   #define MIPI_INTR_STAT(port)		_MMIO_MIPI(port, _MIPIA_INTR_STAT, _MIPIC_INTR_STAT)
> > -#define _MIPIA_INTR_EN			(dev_priv->mipi_mmio_base + 0xb008)
> > -#define _MIPIC_INTR_EN			(dev_priv->mipi_mmio_base + 0xb808)
> > +#define _MIPIA_INTR_EN			(i915->mipi_mmio_base + 0xb008)
> > +#define _MIPIC_INTR_EN			(i915->mipi_mmio_base + 0xb808)
> >   #define MIPI_INTR_EN(port)		_MMIO_MIPI(port, _MIPIA_INTR_EN, _MIPIC_INTR_EN)
> >   #define  TEARING_EFFECT					(1 << 31)
> >   #define  SPL_PKT_SENT_INTERRUPT				(1 << 30)
> > @@ -10518,8 +10518,8 @@ enum skl_power_gate {
> >   #define  RXSOT_SYNC_ERROR				(1 << 1)
> >   #define  RXSOT_ERROR					(1 << 0)
> > -#define _MIPIA_DSI_FUNC_PRG		(dev_priv->mipi_mmio_base + 0xb00c)
> > -#define _MIPIC_DSI_FUNC_PRG		(dev_priv->mipi_mmio_base + 0xb80c)
> > +#define _MIPIA_DSI_FUNC_PRG		(i915->mipi_mmio_base + 0xb00c)
> > +#define _MIPIC_DSI_FUNC_PRG		(i915->mipi_mmio_base + 0xb80c)
> >   #define MIPI_DSI_FUNC_PRG(port)		_MMIO_MIPI(port, _MIPIA_DSI_FUNC_PRG, _MIPIC_DSI_FUNC_PRG)
> >   #define  CMD_MODE_DATA_WIDTH_MASK			(7 << 13)
> >   #define  CMD_MODE_NOT_SUPPORTED				(0 << 13)
> > @@ -10541,78 +10541,78 @@ enum skl_power_gate {
> >   #define  DATA_LANES_PRG_REG_SHIFT			0
> >   #define  DATA_LANES_PRG_REG_MASK			(7 << 0)
> > -#define _MIPIA_HS_TX_TIMEOUT		(dev_priv->mipi_mmio_base + 0xb010)
> > -#define _MIPIC_HS_TX_TIMEOUT		(dev_priv->mipi_mmio_base + 0xb810)
> > +#define _MIPIA_HS_TX_TIMEOUT		(i915->mipi_mmio_base + 0xb010)
> > +#define _MIPIC_HS_TX_TIMEOUT		(i915->mipi_mmio_base + 0xb810)
> >   #define MIPI_HS_TX_TIMEOUT(port)	_MMIO_MIPI(port, _MIPIA_HS_TX_TIMEOUT, _MIPIC_HS_TX_TIMEOUT)
> >   #define  HIGH_SPEED_TX_TIMEOUT_COUNTER_MASK		0xffffff
> > -#define _MIPIA_LP_RX_TIMEOUT		(dev_priv->mipi_mmio_base + 0xb014)
> > -#define _MIPIC_LP_RX_TIMEOUT		(dev_priv->mipi_mmio_base + 0xb814)
> > +#define _MIPIA_LP_RX_TIMEOUT		(i915->mipi_mmio_base + 0xb014)
> > +#define _MIPIC_LP_RX_TIMEOUT		(i915->mipi_mmio_base + 0xb814)
> >   #define MIPI_LP_RX_TIMEOUT(port)	_MMIO_MIPI(port, _MIPIA_LP_RX_TIMEOUT, _MIPIC_LP_RX_TIMEOUT)
> >   #define  LOW_POWER_RX_TIMEOUT_COUNTER_MASK		0xffffff
> > -#define _MIPIA_TURN_AROUND_TIMEOUT	(dev_priv->mipi_mmio_base + 0xb018)
> > -#define _MIPIC_TURN_AROUND_TIMEOUT	(dev_priv->mipi_mmio_base + 0xb818)
> > +#define _MIPIA_TURN_AROUND_TIMEOUT	(i915->mipi_mmio_base + 0xb018)
> > +#define _MIPIC_TURN_AROUND_TIMEOUT	(i915->mipi_mmio_base + 0xb818)
> >   #define MIPI_TURN_AROUND_TIMEOUT(port)	_MMIO_MIPI(port, _MIPIA_TURN_AROUND_TIMEOUT, _MIPIC_TURN_AROUND_TIMEOUT)
> >   #define  TURN_AROUND_TIMEOUT_MASK			0x3f
> > -#define _MIPIA_DEVICE_RESET_TIMER	(dev_priv->mipi_mmio_base + 0xb01c)
> > -#define _MIPIC_DEVICE_RESET_TIMER	(dev_priv->mipi_mmio_base + 0xb81c)
> > +#define _MIPIA_DEVICE_RESET_TIMER	(i915->mipi_mmio_base + 0xb01c)
> > +#define _MIPIC_DEVICE_RESET_TIMER	(i915->mipi_mmio_base + 0xb81c)
> >   #define MIPI_DEVICE_RESET_TIMER(port)	_MMIO_MIPI(port, _MIPIA_DEVICE_RESET_TIMER, _MIPIC_DEVICE_RESET_TIMER)
> >   #define  DEVICE_RESET_TIMER_MASK			0xffff
> > -#define _MIPIA_DPI_RESOLUTION		(dev_priv->mipi_mmio_base + 0xb020)
> > -#define _MIPIC_DPI_RESOLUTION		(dev_priv->mipi_mmio_base + 0xb820)
> > +#define _MIPIA_DPI_RESOLUTION		(i915->mipi_mmio_base + 0xb020)
> > +#define _MIPIC_DPI_RESOLUTION		(i915->mipi_mmio_base + 0xb820)
> >   #define MIPI_DPI_RESOLUTION(port)	_MMIO_MIPI(port, _MIPIA_DPI_RESOLUTION, _MIPIC_DPI_RESOLUTION)
> >   #define  VERTICAL_ADDRESS_SHIFT				16
> >   #define  VERTICAL_ADDRESS_MASK				(0xffff << 16)
> >   #define  HORIZONTAL_ADDRESS_SHIFT			0
> >   #define  HORIZONTAL_ADDRESS_MASK			0xffff
> > -#define _MIPIA_DBI_FIFO_THROTTLE	(dev_priv->mipi_mmio_base + 0xb024)
> > -#define _MIPIC_DBI_FIFO_THROTTLE	(dev_priv->mipi_mmio_base + 0xb824)
> > +#define _MIPIA_DBI_FIFO_THROTTLE	(i915->mipi_mmio_base + 0xb024)
> > +#define _MIPIC_DBI_FIFO_THROTTLE	(i915->mipi_mmio_base + 0xb824)
> >   #define MIPI_DBI_FIFO_THROTTLE(port)	_MMIO_MIPI(port, _MIPIA_DBI_FIFO_THROTTLE, _MIPIC_DBI_FIFO_THROTTLE)
> >   #define  DBI_FIFO_EMPTY_HALF				(0 << 0)
> >   #define  DBI_FIFO_EMPTY_QUARTER				(1 << 0)
> >   #define  DBI_FIFO_EMPTY_7_LOCATIONS			(2 << 0)
> >   /* regs below are bits 15:0 */
> > -#define _MIPIA_HSYNC_PADDING_COUNT	(dev_priv->mipi_mmio_base + 0xb028)
> > -#define _MIPIC_HSYNC_PADDING_COUNT	(dev_priv->mipi_mmio_base + 0xb828)
> > +#define _MIPIA_HSYNC_PADDING_COUNT	(i915->mipi_mmio_base + 0xb028)
> > +#define _MIPIC_HSYNC_PADDING_COUNT	(i915->mipi_mmio_base + 0xb828)
> >   #define MIPI_HSYNC_PADDING_COUNT(port)	_MMIO_MIPI(port, _MIPIA_HSYNC_PADDING_COUNT, _MIPIC_HSYNC_PADDING_COUNT)
> > -#define _MIPIA_HBP_COUNT		(dev_priv->mipi_mmio_base + 0xb02c)
> > -#define _MIPIC_HBP_COUNT		(dev_priv->mipi_mmio_base + 0xb82c)
> > +#define _MIPIA_HBP_COUNT		(i915->mipi_mmio_base + 0xb02c)
> > +#define _MIPIC_HBP_COUNT		(i915->mipi_mmio_base + 0xb82c)
> >   #define MIPI_HBP_COUNT(port)		_MMIO_MIPI(port, _MIPIA_HBP_COUNT, _MIPIC_HBP_COUNT)
> > -#define _MIPIA_HFP_COUNT		(dev_priv->mipi_mmio_base + 0xb030)
> > -#define _MIPIC_HFP_COUNT		(dev_priv->mipi_mmio_base + 0xb830)
> > +#define _MIPIA_HFP_COUNT		(i915->mipi_mmio_base + 0xb030)
> > +#define _MIPIC_HFP_COUNT		(i915->mipi_mmio_base + 0xb830)
> >   #define MIPI_HFP_COUNT(port)		_MMIO_MIPI(port, _MIPIA_HFP_COUNT, _MIPIC_HFP_COUNT)
> > -#define _MIPIA_HACTIVE_AREA_COUNT	(dev_priv->mipi_mmio_base + 0xb034)
> > -#define _MIPIC_HACTIVE_AREA_COUNT	(dev_priv->mipi_mmio_base + 0xb834)
> > +#define _MIPIA_HACTIVE_AREA_COUNT	(i915->mipi_mmio_base + 0xb034)
> > +#define _MIPIC_HACTIVE_AREA_COUNT	(i915->mipi_mmio_base + 0xb834)
> >   #define MIPI_HACTIVE_AREA_COUNT(port)	_MMIO_MIPI(port, _MIPIA_HACTIVE_AREA_COUNT, _MIPIC_HACTIVE_AREA_COUNT)
> > -#define _MIPIA_VSYNC_PADDING_COUNT	(dev_priv->mipi_mmio_base + 0xb038)
> > -#define _MIPIC_VSYNC_PADDING_COUNT	(dev_priv->mipi_mmio_base + 0xb838)
> > +#define _MIPIA_VSYNC_PADDING_COUNT	(i915->mipi_mmio_base + 0xb038)
> > +#define _MIPIC_VSYNC_PADDING_COUNT	(i915->mipi_mmio_base + 0xb838)
> >   #define MIPI_VSYNC_PADDING_COUNT(port)	_MMIO_MIPI(port, _MIPIA_VSYNC_PADDING_COUNT, _MIPIC_VSYNC_PADDING_COUNT)
> > -#define _MIPIA_VBP_COUNT		(dev_priv->mipi_mmio_base + 0xb03c)
> > -#define _MIPIC_VBP_COUNT		(dev_priv->mipi_mmio_base + 0xb83c)
> > +#define _MIPIA_VBP_COUNT		(i915->mipi_mmio_base + 0xb03c)
> > +#define _MIPIC_VBP_COUNT		(i915->mipi_mmio_base + 0xb83c)
> >   #define MIPI_VBP_COUNT(port)		_MMIO_MIPI(port, _MIPIA_VBP_COUNT, _MIPIC_VBP_COUNT)
> > -#define _MIPIA_VFP_COUNT		(dev_priv->mipi_mmio_base + 0xb040)
> > -#define _MIPIC_VFP_COUNT		(dev_priv->mipi_mmio_base + 0xb840)
> > +#define _MIPIA_VFP_COUNT		(i915->mipi_mmio_base + 0xb040)
> > +#define _MIPIC_VFP_COUNT		(i915->mipi_mmio_base + 0xb840)
> >   #define MIPI_VFP_COUNT(port)		_MMIO_MIPI(port, _MIPIA_VFP_COUNT, _MIPIC_VFP_COUNT)
> > -#define _MIPIA_HIGH_LOW_SWITCH_COUNT	(dev_priv->mipi_mmio_base + 0xb044)
> > -#define _MIPIC_HIGH_LOW_SWITCH_COUNT	(dev_priv->mipi_mmio_base + 0xb844)
> > +#define _MIPIA_HIGH_LOW_SWITCH_COUNT	(i915->mipi_mmio_base + 0xb044)
> > +#define _MIPIC_HIGH_LOW_SWITCH_COUNT	(i915->mipi_mmio_base + 0xb844)
> >   #define MIPI_HIGH_LOW_SWITCH_COUNT(port)	_MMIO_MIPI(port,	_MIPIA_HIGH_LOW_SWITCH_COUNT, _MIPIC_HIGH_LOW_SWITCH_COUNT)
> >   /* regs above are bits 15:0 */
> > -#define _MIPIA_DPI_CONTROL		(dev_priv->mipi_mmio_base + 0xb048)
> > -#define _MIPIC_DPI_CONTROL		(dev_priv->mipi_mmio_base + 0xb848)
> > +#define _MIPIA_DPI_CONTROL		(i915->mipi_mmio_base + 0xb048)
> > +#define _MIPIC_DPI_CONTROL		(i915->mipi_mmio_base + 0xb848)
> >   #define MIPI_DPI_CONTROL(port)		_MMIO_MIPI(port, _MIPIA_DPI_CONTROL, _MIPIC_DPI_CONTROL)
> >   #define  DPI_LP_MODE					(1 << 6)
> >   #define  BACKLIGHT_OFF					(1 << 5)
> > @@ -10622,27 +10622,27 @@ enum skl_power_gate {
> >   #define  TURN_ON					(1 << 1)
> >   #define  SHUTDOWN					(1 << 0)
> > -#define _MIPIA_DPI_DATA			(dev_priv->mipi_mmio_base + 0xb04c)
> > -#define _MIPIC_DPI_DATA			(dev_priv->mipi_mmio_base + 0xb84c)
> > +#define _MIPIA_DPI_DATA			(i915->mipi_mmio_base + 0xb04c)
> > +#define _MIPIC_DPI_DATA			(i915->mipi_mmio_base + 0xb84c)
> >   #define MIPI_DPI_DATA(port)		_MMIO_MIPI(port, _MIPIA_DPI_DATA, _MIPIC_DPI_DATA)
> >   #define  COMMAND_BYTE_SHIFT				0
> >   #define  COMMAND_BYTE_MASK				(0x3f << 0)
> > -#define _MIPIA_INIT_COUNT		(dev_priv->mipi_mmio_base + 0xb050)
> > -#define _MIPIC_INIT_COUNT		(dev_priv->mipi_mmio_base + 0xb850)
> > +#define _MIPIA_INIT_COUNT		(i915->mipi_mmio_base + 0xb050)
> > +#define _MIPIC_INIT_COUNT		(i915->mipi_mmio_base + 0xb850)
> >   #define MIPI_INIT_COUNT(port)		_MMIO_MIPI(port, _MIPIA_INIT_COUNT, _MIPIC_INIT_COUNT)
> >   #define  MASTER_INIT_TIMER_SHIFT			0
> >   #define  MASTER_INIT_TIMER_MASK				(0xffff << 0)
> > -#define _MIPIA_MAX_RETURN_PKT_SIZE	(dev_priv->mipi_mmio_base + 0xb054)
> > -#define _MIPIC_MAX_RETURN_PKT_SIZE	(dev_priv->mipi_mmio_base + 0xb854)
> > +#define _MIPIA_MAX_RETURN_PKT_SIZE	(i915->mipi_mmio_base + 0xb054)
> > +#define _MIPIC_MAX_RETURN_PKT_SIZE	(i915->mipi_mmio_base + 0xb854)
> >   #define MIPI_MAX_RETURN_PKT_SIZE(port)	_MMIO_MIPI(port, \
> >   			_MIPIA_MAX_RETURN_PKT_SIZE, _MIPIC_MAX_RETURN_PKT_SIZE)
> >   #define  MAX_RETURN_PKT_SIZE_SHIFT			0
> >   #define  MAX_RETURN_PKT_SIZE_MASK			(0x3ff << 0)
> > -#define _MIPIA_VIDEO_MODE_FORMAT	(dev_priv->mipi_mmio_base + 0xb058)
> > -#define _MIPIC_VIDEO_MODE_FORMAT	(dev_priv->mipi_mmio_base + 0xb858)
> > +#define _MIPIA_VIDEO_MODE_FORMAT	(i915->mipi_mmio_base + 0xb058)
> > +#define _MIPIC_VIDEO_MODE_FORMAT	(i915->mipi_mmio_base + 0xb858)
> >   #define MIPI_VIDEO_MODE_FORMAT(port)	_MMIO_MIPI(port, _MIPIA_VIDEO_MODE_FORMAT, _MIPIC_VIDEO_MODE_FORMAT)
> >   #define  RANDOM_DPI_DISPLAY_RESOLUTION			(1 << 4)
> >   #define  DISABLE_VIDEO_BTA				(1 << 3)
> > @@ -10651,8 +10651,8 @@ enum skl_power_gate {
> >   #define  VIDEO_MODE_NON_BURST_WITH_SYNC_EVENTS		(2 << 0)
> >   #define  VIDEO_MODE_BURST				(3 << 0)
> > -#define _MIPIA_EOT_DISABLE		(dev_priv->mipi_mmio_base + 0xb05c)
> > -#define _MIPIC_EOT_DISABLE		(dev_priv->mipi_mmio_base + 0xb85c)
> > +#define _MIPIA_EOT_DISABLE		(i915->mipi_mmio_base + 0xb05c)
> > +#define _MIPIC_EOT_DISABLE		(i915->mipi_mmio_base + 0xb85c)
> >   #define MIPI_EOT_DISABLE(port)		_MMIO_MIPI(port, _MIPIA_EOT_DISABLE, _MIPIC_EOT_DISABLE)
> >   #define  BXT_DEFEATURE_DPI_FIFO_CTR			(1 << 9)
> >   #define  BXT_DPHY_DEFEATURE_EN				(1 << 8)
> > @@ -10665,35 +10665,35 @@ enum skl_power_gate {
> >   #define  CLOCKSTOP					(1 << 1)
> >   #define  EOT_DISABLE					(1 << 0)
> > -#define _MIPIA_LP_BYTECLK		(dev_priv->mipi_mmio_base + 0xb060)
> > -#define _MIPIC_LP_BYTECLK		(dev_priv->mipi_mmio_base + 0xb860)
> > +#define _MIPIA_LP_BYTECLK		(i915->mipi_mmio_base + 0xb060)
> > +#define _MIPIC_LP_BYTECLK		(i915->mipi_mmio_base + 0xb860)
> >   #define MIPI_LP_BYTECLK(port)		_MMIO_MIPI(port, _MIPIA_LP_BYTECLK, _MIPIC_LP_BYTECLK)
> >   #define  LP_BYTECLK_SHIFT				0
> >   #define  LP_BYTECLK_MASK				(0xffff << 0)
> > -#define _MIPIA_TLPX_TIME_COUNT		(dev_priv->mipi_mmio_base + 0xb0a4)
> > -#define _MIPIC_TLPX_TIME_COUNT		(dev_priv->mipi_mmio_base + 0xb8a4)
> > +#define _MIPIA_TLPX_TIME_COUNT		(i915->mipi_mmio_base + 0xb0a4)
> > +#define _MIPIC_TLPX_TIME_COUNT		(i915->mipi_mmio_base + 0xb8a4)
> >   #define MIPI_TLPX_TIME_COUNT(port)	 _MMIO_MIPI(port, _MIPIA_TLPX_TIME_COUNT, _MIPIC_TLPX_TIME_COUNT)
> > -#define _MIPIA_CLK_LANE_TIMING		(dev_priv->mipi_mmio_base + 0xb098)
> > -#define _MIPIC_CLK_LANE_TIMING		(dev_priv->mipi_mmio_base + 0xb898)
> > +#define _MIPIA_CLK_LANE_TIMING		(i915->mipi_mmio_base + 0xb098)
> > +#define _MIPIC_CLK_LANE_TIMING		(i915->mipi_mmio_base + 0xb898)
> >   #define MIPI_CLK_LANE_TIMING(port)	 _MMIO_MIPI(port, _MIPIA_CLK_LANE_TIMING, _MIPIC_CLK_LANE_TIMING)
> >   /* bits 31:0 */
> > -#define _MIPIA_LP_GEN_DATA		(dev_priv->mipi_mmio_base + 0xb064)
> > -#define _MIPIC_LP_GEN_DATA		(dev_priv->mipi_mmio_base + 0xb864)
> > +#define _MIPIA_LP_GEN_DATA		(i915->mipi_mmio_base + 0xb064)
> > +#define _MIPIC_LP_GEN_DATA		(i915->mipi_mmio_base + 0xb864)
> >   #define MIPI_LP_GEN_DATA(port)		_MMIO_MIPI(port, _MIPIA_LP_GEN_DATA, _MIPIC_LP_GEN_DATA)
> >   /* bits 31:0 */
> > -#define _MIPIA_HS_GEN_DATA		(dev_priv->mipi_mmio_base + 0xb068)
> > -#define _MIPIC_HS_GEN_DATA		(dev_priv->mipi_mmio_base + 0xb868)
> > +#define _MIPIA_HS_GEN_DATA		(i915->mipi_mmio_base + 0xb068)
> > +#define _MIPIC_HS_GEN_DATA		(i915->mipi_mmio_base + 0xb868)
> >   #define MIPI_HS_GEN_DATA(port)		_MMIO_MIPI(port, _MIPIA_HS_GEN_DATA, _MIPIC_HS_GEN_DATA)
> > -#define _MIPIA_LP_GEN_CTRL		(dev_priv->mipi_mmio_base + 0xb06c)
> > -#define _MIPIC_LP_GEN_CTRL		(dev_priv->mipi_mmio_base + 0xb86c)
> > +#define _MIPIA_LP_GEN_CTRL		(i915->mipi_mmio_base + 0xb06c)
> > +#define _MIPIC_LP_GEN_CTRL		(i915->mipi_mmio_base + 0xb86c)
> >   #define MIPI_LP_GEN_CTRL(port)		_MMIO_MIPI(port, _MIPIA_LP_GEN_CTRL, _MIPIC_LP_GEN_CTRL)
> > -#define _MIPIA_HS_GEN_CTRL		(dev_priv->mipi_mmio_base + 0xb070)
> > -#define _MIPIC_HS_GEN_CTRL		(dev_priv->mipi_mmio_base + 0xb870)
> > +#define _MIPIA_HS_GEN_CTRL		(i915->mipi_mmio_base + 0xb070)
> > +#define _MIPIC_HS_GEN_CTRL		(i915->mipi_mmio_base + 0xb870)
> >   #define MIPI_HS_GEN_CTRL(port)		_MMIO_MIPI(port, _MIPIA_HS_GEN_CTRL, _MIPIC_HS_GEN_CTRL)
> >   #define  LONG_PACKET_WORD_COUNT_SHIFT			8
> >   #define  LONG_PACKET_WORD_COUNT_MASK			(0xffff << 8)
> > @@ -10705,8 +10705,8 @@ enum skl_power_gate {
> >   #define  DATA_TYPE_MASK					(0x3f << 0)
> >   /* data type values, see include/video/mipi_display.h */
> > -#define _MIPIA_GEN_FIFO_STAT		(dev_priv->mipi_mmio_base + 0xb074)
> > -#define _MIPIC_GEN_FIFO_STAT		(dev_priv->mipi_mmio_base + 0xb874)
> > +#define _MIPIA_GEN_FIFO_STAT		(i915->mipi_mmio_base + 0xb074)
> > +#define _MIPIC_GEN_FIFO_STAT		(i915->mipi_mmio_base + 0xb874)
> >   #define MIPI_GEN_FIFO_STAT(port)	_MMIO_MIPI(port, _MIPIA_GEN_FIFO_STAT, _MIPIC_GEN_FIFO_STAT)
> >   #define  DPI_FIFO_EMPTY					(1 << 28)
> >   #define  DBI_FIFO_EMPTY					(1 << 27)
> > @@ -10723,15 +10723,15 @@ enum skl_power_gate {
> >   #define  HS_DATA_FIFO_HALF_EMPTY			(1 << 1)
> >   #define  HS_DATA_FIFO_FULL				(1 << 0)
> > -#define _MIPIA_HS_LS_DBI_ENABLE		(dev_priv->mipi_mmio_base + 0xb078)
> > -#define _MIPIC_HS_LS_DBI_ENABLE		(dev_priv->mipi_mmio_base + 0xb878)
> > +#define _MIPIA_HS_LS_DBI_ENABLE		(i915->mipi_mmio_base + 0xb078)
> > +#define _MIPIC_HS_LS_DBI_ENABLE		(i915->mipi_mmio_base + 0xb878)
> >   #define MIPI_HS_LP_DBI_ENABLE(port)	_MMIO_MIPI(port, _MIPIA_HS_LS_DBI_ENABLE, _MIPIC_HS_LS_DBI_ENABLE)
> >   #define  DBI_HS_LP_MODE_MASK				(1 << 0)
> >   #define  DBI_LP_MODE					(1 << 0)
> >   #define  DBI_HS_MODE					(0 << 0)
> > -#define _MIPIA_DPHY_PARAM		(dev_priv->mipi_mmio_base + 0xb080)
> > -#define _MIPIC_DPHY_PARAM		(dev_priv->mipi_mmio_base + 0xb880)
> > +#define _MIPIA_DPHY_PARAM		(i915->mipi_mmio_base + 0xb080)
> > +#define _MIPIC_DPHY_PARAM		(i915->mipi_mmio_base + 0xb880)
> >   #define MIPI_DPHY_PARAM(port)		_MMIO_MIPI(port, _MIPIA_DPHY_PARAM, _MIPIC_DPHY_PARAM)
> >   #define  EXIT_ZERO_COUNT_SHIFT				24
> >   #define  EXIT_ZERO_COUNT_MASK				(0x3f << 24)
> > @@ -10978,34 +10978,34 @@ enum skl_power_gate {
> >   #define  TA_TIMEOUT_VALUE(x)		((x) << 0)
> >   /* bits 31:0 */
> > -#define _MIPIA_DBI_BW_CTRL		(dev_priv->mipi_mmio_base + 0xb084)
> > -#define _MIPIC_DBI_BW_CTRL		(dev_priv->mipi_mmio_base + 0xb884)
> > +#define _MIPIA_DBI_BW_CTRL		(i915->mipi_mmio_base + 0xb084)
> > +#define _MIPIC_DBI_BW_CTRL		(i915->mipi_mmio_base + 0xb884)
> >   #define MIPI_DBI_BW_CTRL(port)		_MMIO_MIPI(port, _MIPIA_DBI_BW_CTRL, _MIPIC_DBI_BW_CTRL)
> > -#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT		(dev_priv->mipi_mmio_base + 0xb088)
> > -#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT		(dev_priv->mipi_mmio_base + 0xb888)
> > +#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT		(i915->mipi_mmio_base + 0xb088)
> > +#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT		(i915->mipi_mmio_base + 0xb888)
> >   #define MIPI_CLK_LANE_SWITCH_TIME_CNT(port)	_MMIO_MIPI(port, _MIPIA_CLK_LANE_SWITCH_TIME_CNT, _MIPIC_CLK_LANE_SWITCH_TIME_CNT)
> >   #define  LP_HS_SSW_CNT_SHIFT				16
> >   #define  LP_HS_SSW_CNT_MASK				(0xffff << 16)
> >   #define  HS_LP_PWR_SW_CNT_SHIFT				0
> >   #define  HS_LP_PWR_SW_CNT_MASK				(0xffff << 0)
> > -#define _MIPIA_STOP_STATE_STALL		(dev_priv->mipi_mmio_base + 0xb08c)
> > -#define _MIPIC_STOP_STATE_STALL		(dev_priv->mipi_mmio_base + 0xb88c)
> > +#define _MIPIA_STOP_STATE_STALL		(i915->mipi_mmio_base + 0xb08c)
> > +#define _MIPIC_STOP_STATE_STALL		(i915->mipi_mmio_base + 0xb88c)
> >   #define MIPI_STOP_STATE_STALL(port)	_MMIO_MIPI(port, _MIPIA_STOP_STATE_STALL, _MIPIC_STOP_STATE_STALL)
> >   #define  STOP_STATE_STALL_COUNTER_SHIFT			0
> >   #define  STOP_STATE_STALL_COUNTER_MASK			(0xff << 0)
> > -#define _MIPIA_INTR_STAT_REG_1		(dev_priv->mipi_mmio_base + 0xb090)
> > -#define _MIPIC_INTR_STAT_REG_1		(dev_priv->mipi_mmio_base + 0xb890)
> > +#define _MIPIA_INTR_STAT_REG_1		(i915->mipi_mmio_base + 0xb090)
> > +#define _MIPIC_INTR_STAT_REG_1		(i915->mipi_mmio_base + 0xb890)
> >   #define MIPI_INTR_STAT_REG_1(port)	_MMIO_MIPI(port, _MIPIA_INTR_STAT_REG_1, _MIPIC_INTR_STAT_REG_1)
> > -#define _MIPIA_INTR_EN_REG_1		(dev_priv->mipi_mmio_base + 0xb094)
> > -#define _MIPIC_INTR_EN_REG_1		(dev_priv->mipi_mmio_base + 0xb894)
> > +#define _MIPIA_INTR_EN_REG_1		(i915->mipi_mmio_base + 0xb094)
> > +#define _MIPIC_INTR_EN_REG_1		(i915->mipi_mmio_base + 0xb894)
> >   #define MIPI_INTR_EN_REG_1(port)	_MMIO_MIPI(port, _MIPIA_INTR_EN_REG_1, _MIPIC_INTR_EN_REG_1)
> >   #define  RX_CONTENTION_DETECTED				(1 << 0)
> >   /* XXX: only pipe A ?!? */
> > -#define MIPIA_DBI_TYPEC_CTRL		(dev_priv->mipi_mmio_base + 0xb100)
> > +#define MIPIA_DBI_TYPEC_CTRL		(i915->mipi_mmio_base + 0xb100)
> >   #define  DBI_TYPEC_ENABLE				(1 << 31)
> >   #define  DBI_TYPEC_WIP					(1 << 30)
> >   #define  DBI_TYPEC_OPTION_SHIFT				28
> > @@ -11019,8 +11019,8 @@ enum skl_power_gate {
> >   /* MIPI adapter registers */
> > -#define _MIPIA_CTRL			(dev_priv->mipi_mmio_base + 0xb104)
> > -#define _MIPIC_CTRL			(dev_priv->mipi_mmio_base + 0xb904)
> > +#define _MIPIA_CTRL			(i915->mipi_mmio_base + 0xb104)
> > +#define _MIPIC_CTRL			(i915->mipi_mmio_base + 0xb904)
> >   #define MIPI_CTRL(port)			_MMIO_MIPI(port, _MIPIA_CTRL, _MIPIC_CTRL)
> >   #define  ESCAPE_CLOCK_DIVIDER_SHIFT			5 /* A only */
> >   #define  ESCAPE_CLOCK_DIVIDER_MASK			(3 << 5)
> > @@ -11052,21 +11052,21 @@ enum skl_power_gate {
> >   #define  GLK_MIPIIO_PORT_POWERED			(1 << 1) /* RO */
> >   #define  GLK_MIPIIO_ENABLE				(1 << 0)
> > -#define _MIPIA_DATA_ADDRESS		(dev_priv->mipi_mmio_base + 0xb108)
> > -#define _MIPIC_DATA_ADDRESS		(dev_priv->mipi_mmio_base + 0xb908)
> > +#define _MIPIA_DATA_ADDRESS		(i915->mipi_mmio_base + 0xb108)
> > +#define _MIPIC_DATA_ADDRESS		(i915->mipi_mmio_base + 0xb908)
> >   #define MIPI_DATA_ADDRESS(port)		_MMIO_MIPI(port, _MIPIA_DATA_ADDRESS, _MIPIC_DATA_ADDRESS)
> >   #define  DATA_MEM_ADDRESS_SHIFT				5
> >   #define  DATA_MEM_ADDRESS_MASK				(0x7ffffff << 5)
> >   #define  DATA_VALID					(1 << 0)
> > -#define _MIPIA_DATA_LENGTH		(dev_priv->mipi_mmio_base + 0xb10c)
> > -#define _MIPIC_DATA_LENGTH		(dev_priv->mipi_mmio_base + 0xb90c)
> > +#define _MIPIA_DATA_LENGTH		(i915->mipi_mmio_base + 0xb10c)
> > +#define _MIPIC_DATA_LENGTH		(i915->mipi_mmio_base + 0xb90c)
> >   #define MIPI_DATA_LENGTH(port)		_MMIO_MIPI(port, _MIPIA_DATA_LENGTH, _MIPIC_DATA_LENGTH)
> >   #define  DATA_LENGTH_SHIFT				0
> >   #define  DATA_LENGTH_MASK				(0xfffff << 0)
> > -#define _MIPIA_COMMAND_ADDRESS		(dev_priv->mipi_mmio_base + 0xb110)
> > -#define _MIPIC_COMMAND_ADDRESS		(dev_priv->mipi_mmio_base + 0xb910)
> > +#define _MIPIA_COMMAND_ADDRESS		(i915->mipi_mmio_base + 0xb110)
> > +#define _MIPIC_COMMAND_ADDRESS		(i915->mipi_mmio_base + 0xb910)
> >   #define MIPI_COMMAND_ADDRESS(port)	_MMIO_MIPI(port, _MIPIA_COMMAND_ADDRESS, _MIPIC_COMMAND_ADDRESS)
> >   #define  COMMAND_MEM_ADDRESS_SHIFT			5
> >   #define  COMMAND_MEM_ADDRESS_MASK			(0x7ffffff << 5)
> > @@ -11074,18 +11074,18 @@ enum skl_power_gate {
> >   #define  MEMORY_WRITE_DATA_FROM_PIPE_RENDERING		(1 << 1)
> >   #define  COMMAND_VALID					(1 << 0)
> > -#define _MIPIA_COMMAND_LENGTH		(dev_priv->mipi_mmio_base + 0xb114)
> > -#define _MIPIC_COMMAND_LENGTH		(dev_priv->mipi_mmio_base + 0xb914)
> > +#define _MIPIA_COMMAND_LENGTH		(i915->mipi_mmio_base + 0xb114)
> > +#define _MIPIC_COMMAND_LENGTH		(i915->mipi_mmio_base + 0xb914)
> >   #define MIPI_COMMAND_LENGTH(port)	_MMIO_MIPI(port, _MIPIA_COMMAND_LENGTH, _MIPIC_COMMAND_LENGTH)
> >   #define  COMMAND_LENGTH_SHIFT(n)			(8 * (n)) /* n: 0...3 */
> >   #define  COMMAND_LENGTH_MASK(n)				(0xff << (8 * (n)))
> > -#define _MIPIA_READ_DATA_RETURN0	(dev_priv->mipi_mmio_base + 0xb118)
> > -#define _MIPIC_READ_DATA_RETURN0	(dev_priv->mipi_mmio_base + 0xb918)
> > +#define _MIPIA_READ_DATA_RETURN0	(i915->mipi_mmio_base + 0xb118)
> > +#define _MIPIC_READ_DATA_RETURN0	(i915->mipi_mmio_base + 0xb918)
> >   #define MIPI_READ_DATA_RETURN(port, n) _MMIO(_MIPI(port, _MIPIA_READ_DATA_RETURN0, _MIPIC_READ_DATA_RETURN0) + 4 * (n)) /* n: 0...7 */
> > -#define _MIPIA_READ_DATA_VALID		(dev_priv->mipi_mmio_base + 0xb138)
> > -#define _MIPIC_READ_DATA_VALID		(dev_priv->mipi_mmio_base + 0xb938)
> > +#define _MIPIA_READ_DATA_VALID		(i915->mipi_mmio_base + 0xb138)
> > +#define _MIPIC_READ_DATA_VALID		(i915->mipi_mmio_base + 0xb938)
> >   #define MIPI_READ_DATA_VALID(port)	_MMIO_MIPI(port, _MIPIA_READ_DATA_VALID, _MIPIC_READ_DATA_VALID)
> >   #define  READ_DATA_VALID(n)				(1 << (n))
> > diff --git a/drivers/gpu/drm/i915/i915_suspend.c b/drivers/gpu/drm/i915/i915_suspend.c
> > index 581201bcb81a..5b3ad7c9240e 100644
> > --- a/drivers/gpu/drm/i915/i915_suspend.c
> > +++ b/drivers/gpu/drm/i915/i915_suspend.c
> > @@ -31,121 +31,121 @@
> >   #include "intel_fbc.h"
> >   #include "intel_gmbus.h"
> > -static void i915_save_display(struct drm_i915_private *dev_priv)
> > +static void i915_save_display(struct drm_i915_private *i915)
> >   {
> >   	/* Display arbitration control */
> > -	if (INTEL_GEN(dev_priv) <= 4)
> > -		dev_priv->regfile.saveDSPARB = I915_READ(DSPARB);
> > +	if (INTEL_GEN(i915) <= 4)
> > +		i915->regfile.saveDSPARB = I915_READ(DSPARB);
> >   	/* save FBC interval */
> > -	if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv))
> > -		dev_priv->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL);
> > +	if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915))
> > +		i915->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL);
> >   }
> > -static void i915_restore_display(struct drm_i915_private *dev_priv)
> > +static void i915_restore_display(struct drm_i915_private *i915)
> >   {
> >   	/* Display arbitration */
> > -	if (INTEL_GEN(dev_priv) <= 4)
> > -		I915_WRITE(DSPARB, dev_priv->regfile.saveDSPARB);
> > +	if (INTEL_GEN(i915) <= 4)
> > +		I915_WRITE(DSPARB, i915->regfile.saveDSPARB);
> >   	/* only restore FBC info on the platform that supports FBC*/
> > -	intel_fbc_global_disable(dev_priv);
> > +	intel_fbc_global_disable(i915);
> >   	/* restore FBC interval */
> > -	if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv))
> > -		I915_WRITE(FBC_CONTROL, dev_priv->regfile.saveFBC_CONTROL);
> > +	if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915))
> > +		I915_WRITE(FBC_CONTROL, i915->regfile.saveFBC_CONTROL);
> > -	i915_redisable_vga(dev_priv);
> > +	i915_redisable_vga(i915);
> >   }
> > -int i915_save_state(struct drm_i915_private *dev_priv)
> > +int i915_save_state(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int i;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > -	i915_save_display(dev_priv);
> > +	i915_save_display(i915);
> > -	if (IS_GEN(dev_priv, 4))
> > +	if (IS_GEN(i915, 4))
> >   		pci_read_config_word(pdev, GCDGMBUS,
> > -				     &dev_priv->regfile.saveGCDGMBUS);
> > +				     &i915->regfile.saveGCDGMBUS);
> >   	/* Cache mode state */
> > -	if (INTEL_GEN(dev_priv) < 7)
> > -		dev_priv->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0);
> > +	if (INTEL_GEN(i915) < 7)
> > +		i915->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0);
> >   	/* Memory Arbitration state */
> > -	dev_priv->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE);
> > +	i915->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE);
> >   	/* Scratch space */
> > -	if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) {
> > +	if (IS_GEN(i915, 2) && IS_MOBILE(i915)) {
> >   		for (i = 0; i < 7; i++) {
> > -			dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> > -			dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> > +			i915->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> > +			i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> >   		}
> >   		for (i = 0; i < 3; i++)
> > -			dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > +			i915->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> > +	} else if (IS_GEN(i915, 2)) {
> >   		for (i = 0; i < 7; i++)
> > -			dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> > -	} else if (HAS_GMCH(dev_priv)) {
> > +			i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> > +	} else if (HAS_GMCH(i915)) {
> >   		for (i = 0; i < 16; i++) {
> > -			dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> > -			dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> > +			i915->regfile.saveSWF0[i] = I915_READ(SWF0(i));
> > +			i915->regfile.saveSWF1[i] = I915_READ(SWF1(i));
> >   		}
> >   		for (i = 0; i < 3; i++)
> > -			dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> > +			i915->regfile.saveSWF3[i] = I915_READ(SWF3(i));
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return 0;
> >   }
> > -int i915_restore_state(struct drm_i915_private *dev_priv)
> > +int i915_restore_state(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	int i;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > -	if (IS_GEN(dev_priv, 4))
> > +	if (IS_GEN(i915, 4))
> >   		pci_write_config_word(pdev, GCDGMBUS,
> > -				      dev_priv->regfile.saveGCDGMBUS);
> > -	i915_restore_display(dev_priv);
> > +				      i915->regfile.saveGCDGMBUS);
> > +	i915_restore_display(i915);
> >   	/* Cache mode state */
> > -	if (INTEL_GEN(dev_priv) < 7)
> > -		I915_WRITE(CACHE_MODE_0, dev_priv->regfile.saveCACHE_MODE_0 |
> > +	if (INTEL_GEN(i915) < 7)
> > +		I915_WRITE(CACHE_MODE_0, i915->regfile.saveCACHE_MODE_0 |
> >   			   0xffff0000);
> >   	/* Memory arbitration state */
> > -	I915_WRITE(MI_ARB_STATE, dev_priv->regfile.saveMI_ARB_STATE | 0xffff0000);
> > +	I915_WRITE(MI_ARB_STATE, i915->regfile.saveMI_ARB_STATE | 0xffff0000);
> >   	/* Scratch space */
> > -	if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) {
> > +	if (IS_GEN(i915, 2) && IS_MOBILE(i915)) {
> >   		for (i = 0; i < 7; i++) {
> > -			I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]);
> > -			I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> > +			I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]);
> > +			I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> >   		}
> >   		for (i = 0; i < 3; i++)
> > -			I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]);
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > +			I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]);
> > +	} else if (IS_GEN(i915, 2)) {
> >   		for (i = 0; i < 7; i++)
> > -			I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> > -	} else if (HAS_GMCH(dev_priv)) {
> > +			I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> > +	} else if (HAS_GMCH(i915)) {
> >   		for (i = 0; i < 16; i++) {
> > -			I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]);
> > -			I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]);
> > +			I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]);
> > +			I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]);
> >   		}
> >   		for (i = 0; i < 3; i++)
> > -			I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]);
> > +			I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]);
> >   	}
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	intel_gmbus_reset(dev_priv);
> > +	intel_gmbus_reset(i915);
> >   	return 0;
> >   }
> > diff --git a/drivers/gpu/drm/i915/i915_sysfs.c b/drivers/gpu/drm/i915/i915_sysfs.c
> > index 3ef07b987d40..9c89be3ec4c1 100644
> > --- a/drivers/gpu/drm/i915/i915_sysfs.c
> > +++ b/drivers/gpu/drm/i915/i915_sysfs.c
> > @@ -42,14 +42,14 @@ static inline struct drm_i915_private *kdev_minor_to_i915(struct device *kdev)
> >   }
> >   #ifdef CONFIG_PM
> > -static u32 calc_residency(struct drm_i915_private *dev_priv,
> > +static u32 calc_residency(struct drm_i915_private *i915,
> >   			  i915_reg_t reg)
> >   {
> >   	intel_wakeref_t wakeref;
> >   	u64 res = 0;
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > -		res = intel_rc6_residency_us(dev_priv, reg);
> > +	with_intel_runtime_pm(i915, wakeref)
> > +		res = intel_rc6_residency_us(i915, reg);
> >   	return DIV_ROUND_CLOSEST_ULL(res, 1000);
> >   }
> > @@ -57,15 +57,15 @@ static u32 calc_residency(struct drm_i915_private *dev_priv,
> >   static ssize_t
> >   show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	unsigned int mask;
> >   	mask = 0;
> > -	if (HAS_RC6(dev_priv))
> > +	if (HAS_RC6(i915))
> >   		mask |= BIT(0);
> > -	if (HAS_RC6p(dev_priv))
> > +	if (HAS_RC6p(i915))
> >   		mask |= BIT(1);
> > -	if (HAS_RC6pp(dev_priv))
> > +	if (HAS_RC6pp(i915))
> >   		mask |= BIT(2);
> >   	return snprintf(buf, PAGE_SIZE, "%x\n", mask);
> > @@ -74,32 +74,32 @@ show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf)
> >   static ssize_t
> >   show_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	u32 rc6_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	u32 rc6_residency = calc_residency(i915, GEN6_GT_GFX_RC6);
> >   	return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency);
> >   }
> >   static ssize_t
> >   show_rc6p_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	u32 rc6p_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6p);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	u32 rc6p_residency = calc_residency(i915, GEN6_GT_GFX_RC6p);
> >   	return snprintf(buf, PAGE_SIZE, "%u\n", rc6p_residency);
> >   }
> >   static ssize_t
> >   show_rc6pp_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	u32 rc6pp_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6pp);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	u32 rc6pp_residency = calc_residency(i915, GEN6_GT_GFX_RC6pp);
> >   	return snprintf(buf, PAGE_SIZE, "%u\n", rc6pp_residency);
> >   }
> >   static ssize_t
> >   show_media_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	u32 rc6_residency = calc_residency(dev_priv, VLV_GT_MEDIA_RC6);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	u32 rc6_residency = calc_residency(i915, VLV_GT_MEDIA_RC6);
> >   	return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency);
> >   }
> > @@ -142,9 +142,9 @@ static const struct attribute_group media_rc6_attr_group = {
> >   };
> >   #endif
> > -static int l3_access_valid(struct drm_i915_private *dev_priv, loff_t offset)
> > +static int l3_access_valid(struct drm_i915_private *i915, loff_t offset)
> >   {
> > -	if (!HAS_L3_DPF(dev_priv))
> > +	if (!HAS_L3_DPF(i915))
> >   		return -EPERM;
> >   	if (offset % 4 != 0)
> > @@ -162,14 +162,14 @@ i915_l3_read(struct file *filp, struct kobject *kobj,
> >   	     loff_t offset, size_t count)
> >   {
> >   	struct device *kdev = kobj_to_dev(kobj);
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct drm_device *dev = &i915->drm;
> >   	int slice = (int)(uintptr_t)attr->private;
> >   	int ret;
> >   	count = round_down(count, 4);
> > -	ret = l3_access_valid(dev_priv, offset);
> > +	ret = l3_access_valid(i915, offset);
> >   	if (ret)
> >   		return ret;
> > @@ -179,9 +179,9 @@ i915_l3_read(struct file *filp, struct kobject *kobj,
> >   	if (ret)
> >   		return ret;
> > -	if (dev_priv->l3_parity.remap_info[slice])
> > +	if (i915->l3_parity.remap_info[slice])
> >   		memcpy(buf,
> > -		       dev_priv->l3_parity.remap_info[slice] + (offset/4),
> > +		       i915->l3_parity.remap_info[slice] + (offset/4),
> >   		       count);
> >   	else
> >   		memset(buf, 0, count);
> > @@ -197,14 +197,14 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> >   	      loff_t offset, size_t count)
> >   {
> >   	struct device *kdev = kobj_to_dev(kobj);
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct i915_gem_context *ctx;
> >   	int slice = (int)(uintptr_t)attr->private;
> >   	u32 **remap_info;
> >   	int ret;
> > -	ret = l3_access_valid(dev_priv, offset);
> > +	ret = l3_access_valid(i915, offset);
> >   	if (ret)
> >   		return ret;
> > @@ -212,7 +212,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> >   	if (ret)
> >   		return ret;
> > -	remap_info = &dev_priv->l3_parity.remap_info[slice];
> > +	remap_info = &i915->l3_parity.remap_info[slice];
> >   	if (!*remap_info) {
> >   		*remap_info = kzalloc(GEN7_L3LOG_SIZE, GFP_KERNEL);
> >   		if (!*remap_info) {
> > @@ -228,7 +228,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj,
> >   	memcpy(*remap_info + (offset/4), buf, count);
> >   	/* NB: We defer the remapping until we switch to the context */
> > -	list_for_each_entry(ctx, &dev_priv->contexts.list, link)
> > +	list_for_each_entry(ctx, &i915->contexts.list, link)
> >   		ctx->remap_slice |= (1<<slice);
> >   	ret = count;
> > @@ -260,52 +260,52 @@ static const struct bin_attribute dpf_attrs_1 = {
> >   static ssize_t gt_act_freq_mhz_show(struct device *kdev,
> >   				    struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	intel_wakeref_t wakeref;
> >   	u32 freq;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		vlv_punit_get(dev_priv);
> > -		freq = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> > -		vlv_punit_put(dev_priv);
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		vlv_punit_get(i915);
> > +		freq = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> > +		vlv_punit_put(i915);
> >   		freq = (freq >> 8) & 0xff;
> >   	} else {
> > -		freq = intel_get_cagf(dev_priv, I915_READ(GEN6_RPSTAT1));
> > +		freq = intel_get_cagf(i915, I915_READ(GEN6_RPSTAT1));
> >   	}
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> > -	return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(dev_priv, freq));
> > +	return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(i915, freq));
> >   }
> >   static ssize_t gt_cur_freq_mhz_show(struct device *kdev,
> >   				    struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	return snprintf(buf, PAGE_SIZE, "%d\n",
> > -			intel_gpu_freq(dev_priv,
> > -				       dev_priv->gt_pm.rps.cur_freq));
> > +			intel_gpu_freq(i915,
> > +				       i915->gt_pm.rps.cur_freq));
> >   }
> >   static ssize_t gt_boost_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	return snprintf(buf, PAGE_SIZE, "%d\n",
> > -			intel_gpu_freq(dev_priv,
> > -				       dev_priv->gt_pm.rps.boost_freq));
> > +			intel_gpu_freq(i915,
> > +				       i915->gt_pm.rps.boost_freq));
> >   }
> >   static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> >   				       struct device_attribute *attr,
> >   				       const char *buf, size_t count)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	bool boost = false;
> >   	ssize_t ret;
> >   	u32 val;
> > @@ -315,7 +315,7 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> >   		return ret;
> >   	/* Validate against (static) hardware limits */
> > -	val = intel_freq_opcode(dev_priv, val);
> > +	val = intel_freq_opcode(i915, val);
> >   	if (val < rps->min_freq || val > rps->max_freq)
> >   		return -EINVAL;
> > @@ -334,28 +334,28 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev,
> >   static ssize_t vlv_rpe_freq_mhz_show(struct device *kdev,
> >   				     struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	return snprintf(buf, PAGE_SIZE, "%d\n",
> > -			intel_gpu_freq(dev_priv,
> > -				       dev_priv->gt_pm.rps.efficient_freq));
> > +			intel_gpu_freq(i915,
> > +				       i915->gt_pm.rps.efficient_freq));
> >   }
> >   static ssize_t gt_max_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	return snprintf(buf, PAGE_SIZE, "%d\n",
> > -			intel_gpu_freq(dev_priv,
> > -				       dev_priv->gt_pm.rps.max_freq_softlimit));
> > +			intel_gpu_freq(i915,
> > +				       i915->gt_pm.rps.max_freq_softlimit));
> >   }
> >   static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> >   				     struct device_attribute *attr,
> >   				     const char *buf, size_t count)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	intel_wakeref_t wakeref;
> >   	u32 val;
> >   	ssize_t ret;
> > @@ -364,10 +364,10 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> >   	if (ret)
> >   		return ret;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	mutex_lock(&rps->lock);
> > -	val = intel_freq_opcode(dev_priv, val);
> > +	val = intel_freq_opcode(i915, val);
> >   	if (val < rps->min_freq ||
> >   	    val > rps->max_freq ||
> >   	    val < rps->min_freq_softlimit) {
> > @@ -377,7 +377,7 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> >   	if (val > rps->rp0_freq)
> >   		DRM_DEBUG("User requested overclocking to %d\n",
> > -			  intel_gpu_freq(dev_priv, val));
> > +			  intel_gpu_freq(i915, val));
> >   	rps->max_freq_softlimit = val;
> > @@ -388,30 +388,30 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev,
> >   	/* We still need *_set_rps to process the new max_delay and
> >   	 * update the interrupt limits and PMINTRMSK even though
> >   	 * frequency request may be unchanged. */
> > -	ret = intel_set_rps(dev_priv, val);
> > +	ret = intel_set_rps(i915, val);
> >   unlock:
> >   	mutex_unlock(&rps->lock);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return ret ?: count;
> >   }
> >   static ssize_t gt_min_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	return snprintf(buf, PAGE_SIZE, "%d\n",
> > -			intel_gpu_freq(dev_priv,
> > -				       dev_priv->gt_pm.rps.min_freq_softlimit));
> > +			intel_gpu_freq(i915,
> > +				       i915->gt_pm.rps.min_freq_softlimit));
> >   }
> >   static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> >   				     struct device_attribute *attr,
> >   				     const char *buf, size_t count)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	intel_wakeref_t wakeref;
> >   	u32 val;
> >   	ssize_t ret;
> > @@ -420,10 +420,10 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> >   	if (ret)
> >   		return ret;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	mutex_lock(&rps->lock);
> > -	val = intel_freq_opcode(dev_priv, val);
> > +	val = intel_freq_opcode(i915, val);
> >   	if (val < rps->min_freq ||
> >   	    val > rps->max_freq ||
> >   	    val > rps->max_freq_softlimit) {
> > @@ -440,11 +440,11 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev,
> >   	/* We still need *_set_rps to process the new min_delay and
> >   	 * update the interrupt limits and PMINTRMSK even though
> >   	 * frequency request may be unchanged. */
> > -	ret = intel_set_rps(dev_priv, val);
> > +	ret = intel_set_rps(i915, val);
> >   unlock:
> >   	mutex_unlock(&rps->lock);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return ret ?: count;
> >   }
> > @@ -465,16 +465,16 @@ static DEVICE_ATTR(gt_RPn_freq_mhz, S_IRUGO, gt_rp_mhz_show, NULL);
> >   /* For now we have a static number of RP states */
> >   static ssize_t gt_rp_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 val;
> >   	if (attr == &dev_attr_gt_RP0_freq_mhz)
> > -		val = intel_gpu_freq(dev_priv, rps->rp0_freq);
> > +		val = intel_gpu_freq(i915, rps->rp0_freq);
> >   	else if (attr == &dev_attr_gt_RP1_freq_mhz)
> > -		val = intel_gpu_freq(dev_priv, rps->rp1_freq);
> > +		val = intel_gpu_freq(i915, rps->rp1_freq);
> >   	else if (attr == &dev_attr_gt_RPn_freq_mhz)
> > -		val = intel_gpu_freq(dev_priv, rps->min_freq);
> > +		val = intel_gpu_freq(i915, rps->min_freq);
> >   	else
> >   		BUG();
> > @@ -540,10 +540,10 @@ static ssize_t error_state_write(struct file *file, struct kobject *kobj,
> >   				 loff_t off, size_t count)
> >   {
> >   	struct device *kdev = kobj_to_dev(kobj);
> > -	struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_minor_to_i915(kdev);
> >   	DRM_DEBUG_DRIVER("Resetting error state\n");
> > -	i915_reset_error_state(dev_priv);
> > +	i915_reset_error_state(i915);
> >   	return count;
> >   }
> > @@ -571,37 +571,37 @@ static void i915_setup_error_capture(struct device *kdev) {}
> >   static void i915_teardown_error_capture(struct device *kdev) {}
> >   #endif
> > -void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> > +void i915_setup_sysfs(struct drm_i915_private *i915)
> >   {
> > -	struct device *kdev = dev_priv->drm.primary->kdev;
> > +	struct device *kdev = i915->drm.primary->kdev;
> >   	int ret;
> >   #ifdef CONFIG_PM
> > -	if (HAS_RC6(dev_priv)) {
> > +	if (HAS_RC6(i915)) {
> >   		ret = sysfs_merge_group(&kdev->kobj,
> >   					&rc6_attr_group);
> >   		if (ret)
> >   			DRM_ERROR("RC6 residency sysfs setup failed\n");
> >   	}
> > -	if (HAS_RC6p(dev_priv)) {
> > +	if (HAS_RC6p(i915)) {
> >   		ret = sysfs_merge_group(&kdev->kobj,
> >   					&rc6p_attr_group);
> >   		if (ret)
> >   			DRM_ERROR("RC6p residency sysfs setup failed\n");
> >   	}
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		ret = sysfs_merge_group(&kdev->kobj,
> >   					&media_rc6_attr_group);
> >   		if (ret)
> >   			DRM_ERROR("Media RC6 residency sysfs setup failed\n");
> >   	}
> >   #endif
> > -	if (HAS_L3_DPF(dev_priv)) {
> > +	if (HAS_L3_DPF(i915)) {
> >   		ret = device_create_bin_file(kdev, &dpf_attrs);
> >   		if (ret)
> >   			DRM_ERROR("l3 parity sysfs setup failed\n");
> > -		if (NUM_L3_SLICES(dev_priv) > 1) {
> > +		if (NUM_L3_SLICES(i915) > 1) {
> >   			ret = device_create_bin_file(kdev,
> >   						     &dpf_attrs_1);
> >   			if (ret)
> > @@ -610,9 +610,9 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> >   	}
> >   	ret = 0;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		ret = sysfs_create_files(&kdev->kobj, vlv_attrs);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > +	else if (INTEL_GEN(i915) >= 6)
> >   		ret = sysfs_create_files(&kdev->kobj, gen6_attrs);
> >   	if (ret)
> >   		DRM_ERROR("RPS sysfs setup failed\n");
> > @@ -620,13 +620,13 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv)
> >   	i915_setup_error_capture(kdev);
> >   }
> > -void i915_teardown_sysfs(struct drm_i915_private *dev_priv)
> > +void i915_teardown_sysfs(struct drm_i915_private *i915)
> >   {
> > -	struct device *kdev = dev_priv->drm.primary->kdev;
> > +	struct device *kdev = i915->drm.primary->kdev;
> >   	i915_teardown_error_capture(kdev);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		sysfs_remove_files(&kdev->kobj, vlv_attrs);
> >   	else
> >   		sysfs_remove_files(&kdev->kobj, gen6_attrs);
> > diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
> > index 5c8cfaa70d72..9548b9fe4199 100644
> > --- a/drivers/gpu/drm/i915/i915_trace.h
> > +++ b/drivers/gpu/drm/i915/i915_trace.h
> > @@ -21,8 +21,8 @@
> >   /* watermark/fifo updates */
> >   TRACE_EVENT(intel_pipe_enable,
> > -	    TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> > -	    TP_ARGS(dev_priv, pipe),
> > +	    TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> > +	    TP_ARGS(i915, pipe),
> >   	    TP_STRUCT__entry(
> >   			     __array(u32, frame, 3)
> > @@ -32,11 +32,11 @@ TRACE_EVENT(intel_pipe_enable,
> >   	    TP_fast_assign(
> >   			   enum pipe _pipe;
> > -			   for_each_pipe(dev_priv, _pipe) {
> > +			   for_each_pipe(i915, _pipe) {
> >   				   __entry->frame[_pipe] =
> > -					   dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe);
> > +					   i915->drm.driver->get_vblank_counter(&i915->drm, _pipe);
> >   				   __entry->scanline[_pipe] =
> > -					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe));
> > +					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe));
> >   			   }
> >   			   __entry->pipe = pipe;
> >   			   ),
> > @@ -49,8 +49,8 @@ TRACE_EVENT(intel_pipe_enable,
> >   );
> >   TRACE_EVENT(intel_pipe_disable,
> > -	    TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> > -	    TP_ARGS(dev_priv, pipe),
> > +	    TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> > +	    TP_ARGS(i915, pipe),
> >   	    TP_STRUCT__entry(
> >   			     __array(u32, frame, 3)
> > @@ -60,11 +60,11 @@ TRACE_EVENT(intel_pipe_disable,
> >   	    TP_fast_assign(
> >   			   enum pipe _pipe;
> > -			   for_each_pipe(dev_priv, _pipe) {
> > +			   for_each_pipe(i915, _pipe) {
> >   				   __entry->frame[_pipe] =
> > -					   dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe);
> > +					   i915->drm.driver->get_vblank_counter(&i915->drm, _pipe);
> >   				   __entry->scanline[_pipe] =
> > -					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe));
> > +					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe));
> >   			   }
> >   			   __entry->pipe = pipe;
> >   			   ),
> > @@ -102,8 +102,8 @@ TRACE_EVENT(intel_pipe_crc,
> >   );
> >   TRACE_EVENT(intel_cpu_fifo_underrun,
> > -	    TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe),
> > -	    TP_ARGS(dev_priv, pipe),
> > +	    TP_PROTO(struct drm_i915_private *i915, enum pipe pipe),
> > +	    TP_ARGS(i915, pipe),
> >   	    TP_STRUCT__entry(
> >   			     __field(enum pipe, pipe)
> > @@ -113,8 +113,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun,
> >   	    TP_fast_assign(
> >   			   __entry->pipe = pipe;
> > -			   __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> > -			   __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> > +			   __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> > +			   __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> >   			   ),
> >   	    TP_printk("pipe %c, frame=%u, scanline=%u",
> > @@ -123,8 +123,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun,
> >   );
> >   TRACE_EVENT(intel_pch_fifo_underrun,
> > -	    TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pch_transcoder),
> > -	    TP_ARGS(dev_priv, pch_transcoder),
> > +	    TP_PROTO(struct drm_i915_private *i915, enum pipe pch_transcoder),
> > +	    TP_ARGS(i915, pch_transcoder),
> >   	    TP_STRUCT__entry(
> >   			     __field(enum pipe, pipe)
> > @@ -135,8 +135,8 @@ TRACE_EVENT(intel_pch_fifo_underrun,
> >   	    TP_fast_assign(
> >   			   enum pipe pipe = pch_transcoder;
> >   			   __entry->pipe = pipe;
> > -			   __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> > -			   __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> > +			   __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> > +			   __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> >   			   ),
> >   	    TP_printk("pch transcoder %c, frame=%u, scanline=%u",
> > @@ -145,8 +145,8 @@ TRACE_EVENT(intel_pch_fifo_underrun,
> >   );
> >   TRACE_EVENT(intel_memory_cxsr,
> > -	    TP_PROTO(struct drm_i915_private *dev_priv, bool old, bool new),
> > -	    TP_ARGS(dev_priv, old, new),
> > +	    TP_PROTO(struct drm_i915_private *i915, bool old, bool new),
> > +	    TP_ARGS(i915, old, new),
> >   	    TP_STRUCT__entry(
> >   			     __array(u32, frame, 3)
> > @@ -157,11 +157,11 @@ TRACE_EVENT(intel_memory_cxsr,
> >   	    TP_fast_assign(
> >   			   enum pipe pipe;
> > -			   for_each_pipe(dev_priv, pipe) {
> > +			   for_each_pipe(i915, pipe) {
> >   				   __entry->frame[pipe] =
> > -					   dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
> > +					   i915->drm.driver->get_vblank_counter(&i915->drm, pipe);
> >   				   __entry->scanline[pipe] =
> > -					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
> > +					   intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe));
> >   			   }
> >   			   __entry->old = old;
> >   			   __entry->new = new;
> > diff --git a/drivers/gpu/drm/i915/i915_vgpu.c b/drivers/gpu/drm/i915/i915_vgpu.c
> > index 94d3992b599d..96874a4819a9 100644
> > --- a/drivers/gpu/drm/i915/i915_vgpu.c
> > +++ b/drivers/gpu/drm/i915/i915_vgpu.c
> > @@ -53,14 +53,14 @@
> >   /**
> >    * i915_check_vgpu - detect virtual GPU
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    *
> >    * This function is called at the initialization stage, to detect whether
> >    * running on a vGPU.
> >    */
> > -void i915_check_vgpu(struct drm_i915_private *dev_priv)
> > +void i915_check_vgpu(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u64 magic;
> >   	u16 version_major;
> > @@ -76,15 +76,15 @@ void i915_check_vgpu(struct drm_i915_private *dev_priv)
> >   		return;
> >   	}
> > -	dev_priv->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps));
> > +	i915->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps));
> > -	dev_priv->vgpu.active = true;
> > +	i915->vgpu.active = true;
> >   	DRM_INFO("Virtual GPU for Intel GVT-g detected.\n");
> >   }
> > -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv)
> > +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->vgpu.caps & VGT_CAPS_FULL_PPGTT;
> > +	return i915->vgpu.caps & VGT_CAPS_FULL_PPGTT;
> >   }
> >   struct _balloon_info_ {
> > @@ -112,22 +112,22 @@ static void vgt_deballoon_space(struct i915_ggtt *ggtt,
> >   /**
> >    * intel_vgt_deballoon - deballoon reserved graphics address trunks
> > - * @dev_priv: i915 device private data
> > + * @i915: i915 device private data
> >    *
> >    * This function is called to deallocate the ballooned-out graphic memory, when
> >    * driver is unloaded or when ballooning fails.
> >    */
> > -void intel_vgt_deballoon(struct drm_i915_private *dev_priv)
> > +void intel_vgt_deballoon(struct drm_i915_private *i915)
> >   {
> >   	int i;
> > -	if (!intel_vgpu_active(dev_priv))
> > +	if (!intel_vgpu_active(i915))
> >   		return;
> >   	DRM_DEBUG("VGT deballoon.\n");
> >   	for (i = 0; i < 4; i++)
> > -		vgt_deballoon_space(&dev_priv->ggtt, &bl_info.space[i]);
> > +		vgt_deballoon_space(&i915->ggtt, &bl_info.space[i]);
> >   }
> >   static int vgt_balloon_space(struct i915_ggtt *ggtt,
> > @@ -153,7 +153,7 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
> >   /**
> >    * intel_vgt_balloon - balloon out reserved graphics address trunks
> > - * @dev_priv: i915 device private data
> > + * @i915: i915 device private data
> >    *
> >    * This function is called at the initialization stage, to balloon out the
> >    * graphic address space allocated to other vGPUs, by marking these spaces as
> > @@ -195,16 +195,16 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt,
> >    * Returns:
> >    * zero on success, non-zero if configuration invalid or ballooning failed
> >    */
> > -int intel_vgt_balloon(struct drm_i915_private *dev_priv)
> > +int intel_vgt_balloon(struct drm_i915_private *i915)
> >   {
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	unsigned long ggtt_end = ggtt->vm.total;
> >   	unsigned long mappable_base, mappable_size, mappable_end;
> >   	unsigned long unmappable_base, unmappable_size, unmappable_end;
> >   	int ret;
> > -	if (!intel_vgpu_active(dev_priv))
> > +	if (!intel_vgpu_active(i915))
> >   		return 0;
> >   	mappable_base = I915_READ(vgtif_reg(avail_rs.mappable_gmadr.base));
> > diff --git a/drivers/gpu/drm/i915/i915_vgpu.h b/drivers/gpu/drm/i915/i915_vgpu.h
> > index ebe1b7bced98..0fa0bfb88bed 100644
> > --- a/drivers/gpu/drm/i915/i915_vgpu.h
> > +++ b/drivers/gpu/drm/i915/i915_vgpu.h
> > @@ -26,23 +26,23 @@
> >   #include "i915_pvinfo.h"
> > -void i915_check_vgpu(struct drm_i915_private *dev_priv);
> > +void i915_check_vgpu(struct drm_i915_private *i915);
> > -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv);
> > +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915);
> >   static inline bool
> > -intel_vgpu_has_hwsp_emulation(struct drm_i915_private *dev_priv)
> > +intel_vgpu_has_hwsp_emulation(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->vgpu.caps & VGT_CAPS_HWSP_EMULATION;
> > +	return i915->vgpu.caps & VGT_CAPS_HWSP_EMULATION;
> >   }
> >   static inline bool
> > -intel_vgpu_has_huge_gtt(struct drm_i915_private *dev_priv)
> > +intel_vgpu_has_huge_gtt(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->vgpu.caps & VGT_CAPS_HUGE_GTT;
> > +	return i915->vgpu.caps & VGT_CAPS_HUGE_GTT;
> >   }
> > -int intel_vgt_balloon(struct drm_i915_private *dev_priv);
> > -void intel_vgt_deballoon(struct drm_i915_private *dev_priv);
> > +int intel_vgt_balloon(struct drm_i915_private *i915);
> > +void intel_vgt_deballoon(struct drm_i915_private *i915);
> >   #endif /* _I915_VGPU_H_ */
> > diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
> > index cb341e4acf99..e3f4d7006a1b 100644
> > --- a/drivers/gpu/drm/i915/i915_vma.c
> > +++ b/drivers/gpu/drm/i915/i915_vma.c
> > @@ -555,7 +555,7 @@ static void assert_bind_count(const struct drm_i915_gem_object *obj)
> >   static int
> >   i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags)
> >   {
> > -	struct drm_i915_private *dev_priv = vma->vm->i915;
> > +	struct drm_i915_private *i915 = vma->vm->i915;
> >   	unsigned int cache_level;
> >   	u64 start, end;
> >   	int ret;
> > @@ -581,7 +581,7 @@ i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags)
> >   	end = vma->vm->total;
> >   	if (flags & PIN_MAPPABLE)
> > -		end = min_t(u64, end, dev_priv->ggtt.mappable_end);
> > +		end = min_t(u64, end, i915->ggtt.mappable_end);
> >   	if (flags & PIN_ZONE_4G)
> >   		end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE);
> >   	GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE));
> > diff --git a/drivers/gpu/drm/i915/icl_dsi.c b/drivers/gpu/drm/i915/icl_dsi.c
> > index 74448e6bf749..d0192e093f6f 100644
> > --- a/drivers/gpu/drm/i915/icl_dsi.c
> > +++ b/drivers/gpu/drm/i915/icl_dsi.c
> > @@ -35,32 +35,32 @@
> >   #include "intel_dsi.h"
> >   #include "intel_panel.h"
> > -static inline int header_credits_available(struct drm_i915_private *dev_priv,
> > +static inline int header_credits_available(struct drm_i915_private *i915,
> >   					   enum transcoder dsi_trans)
> >   {
> >   	return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_HEADER_CREDIT_MASK)
> >   		>> FREE_HEADER_CREDIT_SHIFT;
> >   }
> > -static inline int payload_credits_available(struct drm_i915_private *dev_priv,
> > +static inline int payload_credits_available(struct drm_i915_private *i915,
> >   					    enum transcoder dsi_trans)
> >   {
> >   	return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_PLOAD_CREDIT_MASK)
> >   		>> FREE_PLOAD_CREDIT_SHIFT;
> >   }
> > -static void wait_for_header_credits(struct drm_i915_private *dev_priv,
> > +static void wait_for_header_credits(struct drm_i915_private *i915,
> >   				    enum transcoder dsi_trans)
> >   {
> > -	if (wait_for_us(header_credits_available(dev_priv, dsi_trans) >=
> > +	if (wait_for_us(header_credits_available(i915, dsi_trans) >=
> >   			MAX_HEADER_CREDIT, 100))
> >   		DRM_ERROR("DSI header credits not released\n");
> >   }
> > -static void wait_for_payload_credits(struct drm_i915_private *dev_priv,
> > +static void wait_for_payload_credits(struct drm_i915_private *i915,
> >   				     enum transcoder dsi_trans)
> >   {
> > -	if (wait_for_us(payload_credits_available(dev_priv, dsi_trans) >=
> > +	if (wait_for_us(payload_credits_available(i915, dsi_trans) >=
> >   			MAX_PLOAD_CREDIT, 100))
> >   		DRM_ERROR("DSI payload credits not released\n");
> >   }
> > @@ -75,7 +75,7 @@ static enum transcoder dsi_port_to_transcoder(enum port port)
> >   static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	struct mipi_dsi_device *dsi;
> >   	enum port port;
> > @@ -85,8 +85,8 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> >   	/* wait for header/payload credits to be released */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		dsi_trans = dsi_port_to_transcoder(port);
> > -		wait_for_header_credits(dev_priv, dsi_trans);
> > -		wait_for_payload_credits(dev_priv, dsi_trans);
> > +		wait_for_header_credits(i915, dsi_trans);
> > +		wait_for_payload_credits(i915, dsi_trans);
> >   	}
> >   	/* send nop DCS command */
> > @@ -102,7 +102,7 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder)
> >   	/* wait for header credits to be released */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		dsi_trans = dsi_port_to_transcoder(port);
> > -		wait_for_header_credits(dev_priv, dsi_trans);
> > +		wait_for_header_credits(i915, dsi_trans);
> >   	}
> >   	/* wait for LP TX in progress bit to be cleared */
> > @@ -118,7 +118,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data,
> >   			       u32 len)
> >   {
> >   	struct intel_dsi *intel_dsi = host->intel_dsi;
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> >   	enum transcoder dsi_trans = dsi_port_to_transcoder(host->port);
> >   	int free_credits;
> >   	int i, j;
> > @@ -126,7 +126,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data,
> >   	for (i = 0; i < len; i += 4) {
> >   		u32 tmp = 0;
> > -		free_credits = payload_credits_available(dev_priv, dsi_trans);
> > +		free_credits = payload_credits_available(i915, dsi_trans);
> >   		if (free_credits < 1) {
> >   			DRM_ERROR("Payload credit not available\n");
> >   			return false;
> > @@ -145,13 +145,13 @@ static int dsi_send_pkt_hdr(struct intel_dsi_host *host,
> >   			    struct mipi_dsi_packet pkt, bool enable_lpdt)
> >   {
> >   	struct intel_dsi *intel_dsi = host->intel_dsi;
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> >   	enum transcoder dsi_trans = dsi_port_to_transcoder(host->port);
> >   	u32 tmp;
> >   	int free_credits;
> >   	/* check if header credit available */
> > -	free_credits = header_credits_available(dev_priv, dsi_trans);
> > +	free_credits = header_credits_available(i915, dsi_trans);
> >   	if (free_credits < 1) {
> >   		DRM_ERROR("send pkt header failed, not enough hdr credits\n");
> >   		return -1;
> > @@ -200,7 +200,7 @@ static int dsi_send_pkt_payld(struct intel_dsi_host *host,
> >   static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -266,7 +266,7 @@ static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder)
> >   static void configure_dual_link_mode(struct intel_encoder *encoder,
> >   				     const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 dss_ctl1;
> > @@ -304,7 +304,7 @@ static void configure_dual_link_mode(struct intel_encoder *encoder,
> >   static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> > @@ -329,7 +329,7 @@ static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder)
> >   	}
> >   }
> > -static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
> > +static void get_dsi_io_power_domains(struct drm_i915_private *i915,
> >   				     struct intel_dsi *intel_dsi)
> >   {
> >   	enum port port;
> > @@ -337,7 +337,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		WARN_ON(intel_dsi->io_wakeref[port]);
> >   		intel_dsi->io_wakeref[port] =
> > -			intel_display_power_get(dev_priv,
> > +			intel_display_power_get(i915,
> >   						port == PORT_A ?
> >   						POWER_DOMAIN_PORT_DDI_A_IO :
> >   						POWER_DOMAIN_PORT_DDI_B_IO);
> > @@ -346,7 +346,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv,
> >   static void gen11_dsi_enable_io_power(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -357,23 +357,23 @@ static void gen11_dsi_enable_io_power(struct intel_encoder *encoder)
> >   		I915_WRITE(ICL_DSI_IO_MODECTL(port), tmp);
> >   	}
> > -	get_dsi_io_power_domains(dev_priv, intel_dsi);
> > +	get_dsi_io_power_domains(i915, intel_dsi);
> >   }
> >   static void gen11_dsi_power_up_lanes(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	for_each_dsi_port(port, intel_dsi->ports)
> > -		intel_combo_phy_power_up_lanes(dev_priv, port, true,
> > +		intel_combo_phy_power_up_lanes(i915, port, true,
> >   					       intel_dsi->lane_count, false);
> >   }
> >   static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -409,7 +409,7 @@ static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder)
> >   static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > @@ -461,7 +461,7 @@ static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder)
> >   static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > @@ -480,7 +480,7 @@ static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder)
> >   static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > @@ -535,48 +535,48 @@ static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder)
> >   static void gen11_dsi_gate_clocks(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	tmp = I915_READ(DPCLKA_CFGCR0_ICL);
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		tmp |= DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> >   	}
> >   	I915_WRITE(DPCLKA_CFGCR0_ICL, tmp);
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static void gen11_dsi_ungate_clocks(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	tmp = I915_READ(DPCLKA_CFGCR0_ICL);
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		tmp &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> >   	}
> >   	I915_WRITE(DPCLKA_CFGCR0_ICL, tmp);
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static void gen11_dsi_map_pll(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> >   	enum port port;
> >   	u32 val;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	val = I915_READ(DPCLKA_CFGCR0_ICL);
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > @@ -592,14 +592,14 @@ static void gen11_dsi_map_pll(struct intel_encoder *encoder,
> >   	POSTING_READ(DPCLKA_CFGCR0_ICL);
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static void
> >   gen11_dsi_configure_transcoder(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	enum pipe pipe = intel_crtc->pipe;
> > @@ -736,7 +736,7 @@ static void
> >   gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder,
> >   				 const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	const struct drm_display_mode *adjusted_mode =
> >   					&pipe_config->base.adjusted_mode;
> > @@ -849,7 +849,7 @@ gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder,
> >   static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	enum transcoder dsi_trans;
> > @@ -862,7 +862,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
> >   		I915_WRITE(PIPECONF(dsi_trans), tmp);
> >   		/* wait for transcoder to be enabled */
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    PIPECONF(dsi_trans),
> >   					    I965_PIPECONF_ACTIVE,
> >   					    I965_PIPECONF_ACTIVE, 10))
> > @@ -872,7 +872,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder)
> >   static void gen11_dsi_setup_timeouts(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	enum transcoder dsi_trans;
> > @@ -950,7 +950,7 @@ gen11_dsi_enable_port_and_phy(struct intel_encoder *encoder,
> >   static void gen11_dsi_powerup_panel(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	struct mipi_dsi_device *dsi;
> >   	enum port port;
> > @@ -1026,7 +1026,7 @@ static void gen11_dsi_pre_enable(struct intel_encoder *encoder,
> >   static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	enum transcoder dsi_trans;
> > @@ -1041,7 +1041,7 @@ static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder)
> >   		I915_WRITE(PIPECONF(dsi_trans), tmp);
> >   		/* wait for transcoder to be disabled */
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    PIPECONF(dsi_trans),
> >   					    I965_PIPECONF_ACTIVE, 0, 50))
> >   			DRM_ERROR("DSI trancoder not disabled\n");
> > @@ -1062,7 +1062,7 @@ static void gen11_dsi_powerdown_panel(struct intel_encoder *encoder)
> >   static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	enum transcoder dsi_trans;
> > @@ -1103,7 +1103,7 @@ static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder)
> >   static void gen11_dsi_disable_port(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u32 tmp;
> >   	enum port port;
> > @@ -1125,7 +1125,7 @@ static void gen11_dsi_disable_port(struct intel_encoder *encoder)
> >   static void gen11_dsi_disable_io_power(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -1134,7 +1134,7 @@ static void gen11_dsi_disable_io_power(struct intel_encoder *encoder)
> >   		intel_wakeref_t wakeref;
> >   		wakeref = fetch_and_zero(&intel_dsi->io_wakeref[port]);
> > -		intel_display_power_put(dev_priv,
> > +		intel_display_power_put(i915,
> >   					port == PORT_A ?
> >   					POWER_DOMAIN_PORT_DDI_A_IO :
> >   					POWER_DOMAIN_PORT_DDI_B_IO,
> > @@ -1205,13 +1205,13 @@ static void gen11_dsi_get_timings(struct intel_encoder *encoder,
> >   static void gen11_dsi_get_config(struct intel_encoder *encoder,
> >   				 struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	/* FIXME: adapt icl_ddi_clock_get() for DSI and use that? */
> >   	pipe_config->port_clock =
> > -		cnl_calc_wrpll_link(dev_priv, &pipe_config->dpll_hw_state);
> > +		cnl_calc_wrpll_link(i915, &pipe_config->dpll_hw_state);
> >   	pipe_config->base.adjusted_mode.crtc_clock = intel_dsi->pclk;
> >   	if (intel_dsi->dual_link)
> > @@ -1263,7 +1263,7 @@ static void gen11_dsi_get_power_domains(struct intel_encoder *encoder,
> >   static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> >   				   enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum transcoder dsi_trans;
> >   	intel_wakeref_t wakeref;
> > @@ -1271,7 +1271,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> >   	bool ret = false;
> >   	u32 tmp;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1298,7 +1298,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder,
> >   		ret = tmp & PIPECONF_ENABLE;
> >   	}
> >   out:
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -1391,8 +1391,8 @@ static const struct mipi_dsi_host_ops gen11_dsi_host_ops = {
> >   static void icl_dphy_param_init(struct intel_dsi *intel_dsi)
> >   {
> >   	struct drm_device *dev = intel_dsi->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct mipi_config *mipi_config = i915->vbt.dsi.config;
> >   	u32 tlpx_ns;
> >   	u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt;
> >   	u32 ths_prepare_ns, tclk_trail_ns;
> > @@ -1487,9 +1487,9 @@ static void icl_dphy_param_init(struct intel_dsi *intel_dsi)
> >   	intel_dsi_log_params(intel_dsi);
> >   }
> > -void icl_dsi_init(struct drm_i915_private *dev_priv)
> > +void icl_dsi_init(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_dsi *intel_dsi;
> >   	struct intel_encoder *encoder;
> >   	struct intel_connector *intel_connector;
> > @@ -1497,7 +1497,7 @@ void icl_dsi_init(struct drm_i915_private *dev_priv)
> >   	struct drm_display_mode *fixed_mode;
> >   	enum port port;
> > -	if (!intel_bios_is_dsi_present(dev_priv, &port))
> > +	if (!intel_bios_is_dsi_present(i915, &port))
> >   		return;
> >   	intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL);
> > @@ -1556,13 +1556,13 @@ void icl_dsi_init(struct drm_i915_private *dev_priv)
> >   	intel_panel_init(&intel_connector->panel, fixed_mode, NULL);
> >   	intel_panel_setup_backlight(connector, INVALID_PIPE);
> > -	if (dev_priv->vbt.dsi.config->dual_link)
> > +	if (i915->vbt.dsi.config->dual_link)
> >   		intel_dsi->ports = BIT(PORT_A) | BIT(PORT_B);
> >   	else
> >   		intel_dsi->ports = BIT(port);
> > -	intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports;
> > -	intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports;
> > +	intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports;
> > +	intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports;
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		struct intel_dsi_host *host;
> > diff --git a/drivers/gpu/drm/i915/intel_atomic.c b/drivers/gpu/drm/i915/intel_atomic.c
> > index 6b985e895a97..45782e6ff0f4 100644
> > --- a/drivers/gpu/drm/i915/intel_atomic.c
> > +++ b/drivers/gpu/drm/i915/intel_atomic.c
> > @@ -54,13 +54,13 @@ int intel_digital_connector_atomic_get_property(struct drm_connector *connector,
> >   						u64 *val)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_digital_connector_state *intel_conn_state =
> >   		to_intel_digital_connector_state(state);
> > -	if (property == dev_priv->force_audio_property)
> > +	if (property == i915->force_audio_property)
> >   		*val = intel_conn_state->force_audio;
> > -	else if (property == dev_priv->broadcast_rgb_property)
> > +	else if (property == i915->broadcast_rgb_property)
> >   		*val = intel_conn_state->broadcast_rgb;
> >   	else {
> >   		DRM_DEBUG_ATOMIC("Unknown property [PROP:%d:%s]\n",
> > @@ -86,16 +86,16 @@ int intel_digital_connector_atomic_set_property(struct drm_connector *connector,
> >   						u64 val)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_digital_connector_state *intel_conn_state =
> >   		to_intel_digital_connector_state(state);
> > -	if (property == dev_priv->force_audio_property) {
> > +	if (property == i915->force_audio_property) {
> >   		intel_conn_state->force_audio = val;
> >   		return 0;
> >   	}
> > -	if (property == dev_priv->broadcast_rgb_property) {
> > +	if (property == i915->broadcast_rgb_property) {
> >   		intel_conn_state->broadcast_rgb = val;
> >   		return 0;
> >   	}
> > @@ -227,7 +227,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> >   				      struct intel_plane_state *plane_state,
> >   				      int *scaler_id)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	int j;
> >   	u32 mode;
> > @@ -251,10 +251,10 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> >   	    plane_state->base.fb->format->is_yuv &&
> >   	    plane_state->base.fb->format->num_planes > 1) {
> >   		struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -		if (IS_GEN(dev_priv, 9) &&
> > -		    !IS_GEMINILAKE(dev_priv)) {
> > +		if (IS_GEN(i915, 9) &&
> > +		    !IS_GEMINILAKE(i915)) {
> >   			mode = SKL_PS_SCALER_MODE_NV12;
> > -		} else if (icl_is_hdr_plane(dev_priv, plane->id)) {
> > +		} else if (icl_is_hdr_plane(i915, plane->id)) {
> >   			/*
> >   			 * On gen11+'s HDR planes we only use the scaler for
> >   			 * scaling. They have a dedicated chroma upsampler, so
> > @@ -267,7 +267,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> >   			if (plane_state->linked_plane)
> >   				mode |= PS_PLANE_Y_SEL(plane_state->linked_plane->id);
> >   		}
> > -	} else if (INTEL_GEN(dev_priv) > 9 || IS_GEMINILAKE(dev_priv)) {
> > +	} else if (INTEL_GEN(i915) > 9 || IS_GEMINILAKE(i915)) {
> >   		mode = PS_SCALER_MODE_NORMAL;
> >   	} else if (num_scalers_need == 1 && intel_crtc->num_scalers > 1) {
> >   		/*
> > @@ -290,7 +290,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> >   /**
> >    * intel_atomic_setup_scalers() - setup scalers for crtc per staged requests
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @intel_crtc: intel crtc
> >    * @crtc_state: incoming crtc_state to validate and setup scalers
> >    *
> > @@ -305,7 +305,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta
> >    *         0 - scalers were setup succesfully
> >    *         error code - otherwise
> >    */
> > -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> > +int intel_atomic_setup_scalers(struct drm_i915_private *i915,
> >   			       struct intel_crtc *intel_crtc,
> >   			       struct intel_crtc_state *crtc_state)
> >   {
> > @@ -369,7 +369,7 @@ int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> >   			 */
> >   			if (!plane) {
> >   				struct drm_plane_state *state;
> > -				plane = drm_plane_from_index(&dev_priv->drm, i);
> > +				plane = drm_plane_from_index(&i915->drm, i);
> >   				state = drm_atomic_get_plane_state(drm_state, plane);
> >   				if (IS_ERR(state)) {
> >   					DRM_DEBUG_KMS("Failed to add [PLANE:%d] to drm_state\n",
> > diff --git a/drivers/gpu/drm/i915/intel_atomic.h b/drivers/gpu/drm/i915/intel_atomic.h
> > index 1c8507da1a69..5482fd04d639 100644
> > --- a/drivers/gpu/drm/i915/intel_atomic.h
> > +++ b/drivers/gpu/drm/i915/intel_atomic.h
> > @@ -42,7 +42,7 @@ struct intel_crtc_state *
> >   intel_atomic_get_crtc_state(struct drm_atomic_state *state,
> >   			    struct intel_crtc *crtc);
> > -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv,
> > +int intel_atomic_setup_scalers(struct drm_i915_private *i915,
> >   			       struct intel_crtc *intel_crtc,
> >   			       struct intel_crtc_state *crtc_state);
> > diff --git a/drivers/gpu/drm/i915/intel_audio.c b/drivers/gpu/drm/i915/intel_audio.c
> > index 840daff12246..f463d8452fe9 100644
> > --- a/drivers/gpu/drm/i915/intel_audio.c
> > +++ b/drivers/gpu/drm/i915/intel_audio.c
> > @@ -230,7 +230,7 @@ static bool intel_eld_uptodate(struct drm_connector *connector,
> >   			       i915_reg_t reg_elda, u32 bits_elda,
> >   			       i915_reg_t reg_edid)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	const u8 *eld = connector->eld;
> >   	u32 tmp;
> >   	int i;
> > @@ -256,7 +256,7 @@ static void g4x_audio_codec_disable(struct intel_encoder *encoder,
> >   				    const struct intel_crtc_state *old_crtc_state,
> >   				    const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 eldv, tmp;
> >   	DRM_DEBUG_KMS("Disable audio codec\n");
> > @@ -277,7 +277,7 @@ static void g4x_audio_codec_enable(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *crtc_state,
> >   				   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_connector *connector = conn_state->connector;
> >   	const u8 *eld = connector->eld;
> >   	u32 eldv;
> > @@ -317,8 +317,8 @@ static void
> >   hsw_dp_audio_config_update(struct intel_encoder *encoder,
> >   			   const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	struct i915_audio_component *acomp = dev_priv->audio_component;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	struct i915_audio_component *acomp = i915->audio_component;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	enum port port = encoder->port;
> >   	const struct dp_aud_n_m *nm;
> > @@ -364,8 +364,8 @@ static void
> >   hsw_hdmi_audio_config_update(struct intel_encoder *encoder,
> >   			     const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	struct i915_audio_component *acomp = dev_priv->audio_component;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	struct i915_audio_component *acomp = i915->audio_component;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	enum port port = encoder->port;
> >   	int n, rate;
> > @@ -416,14 +416,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder,
> >   				    const struct intel_crtc_state *old_crtc_state,
> >   				    const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> >   	u32 tmp;
> >   	DRM_DEBUG_KMS("Disable audio codec on transcoder %s\n",
> >   		      transcoder_name(cpu_transcoder));
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> >   	/* Disable timestamps */
> >   	tmp = I915_READ(HSW_AUD_CFG(cpu_transcoder));
> > @@ -441,14 +441,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder,
> >   	tmp &= ~AUDIO_OUTPUT_ENABLE(cpu_transcoder);
> >   	I915_WRITE(HSW_AUD_PIN_ELD_CP_VLD, tmp);
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	mutex_unlock(&i915->av_mutex);
> >   }
> >   static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *crtc_state,
> >   				   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_connector *connector = conn_state->connector;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	const u8 *eld = connector->eld;
> > @@ -458,7 +458,7 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> >   	DRM_DEBUG_KMS("Enable audio codec on transcoder %s, %u bytes ELD\n",
> >   		      transcoder_name(cpu_transcoder), drm_eld_size(eld));
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> >   	/* Enable audio presence detect, invalidate ELD */
> >   	tmp = I915_READ(HSW_AUD_PIN_ELD_CP_VLD);
> > @@ -491,14 +491,14 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder,
> >   	/* Enable timestamps */
> >   	hsw_audio_config_update(encoder, crtc_state);
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	mutex_unlock(&i915->av_mutex);
> >   }
> >   static void ilk_audio_codec_disable(struct intel_encoder *encoder,
> >   				    const struct intel_crtc_state *old_crtc_state,
> >   				    const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum port port = encoder->port;
> > @@ -511,10 +511,10 @@ static void ilk_audio_codec_disable(struct intel_encoder *encoder,
> >   	if (WARN_ON(port == PORT_A))
> >   		return;
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915)) {
> >   		aud_config = IBX_AUD_CFG(pipe);
> >   		aud_cntrl_st2 = IBX_AUD_CNTL_ST2;
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		aud_config = VLV_AUD_CFG(pipe);
> >   		aud_cntrl_st2 = VLV_AUD_CNTL_ST2;
> >   	} else {
> > @@ -544,7 +544,7 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *crtc_state,
> >   				   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct drm_connector *connector = conn_state->connector;
> >   	enum pipe pipe = crtc->pipe;
> > @@ -567,13 +567,13 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder,
> >   	 * infrastructure is not there yet.
> >   	 */
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915)) {
> >   		hdmiw_hdmiedid = IBX_HDMIW_HDMIEDID(pipe);
> >   		aud_config = IBX_AUD_CFG(pipe);
> >   		aud_cntl_st = IBX_AUD_CNTL_ST(pipe);
> >   		aud_cntrl_st2 = IBX_AUD_CNTL_ST2;
> > -	} else if (IS_VALLEYVIEW(dev_priv) ||
> > -		   IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) ||
> > +		   IS_CHERRYVIEW(i915)) {
> >   		hdmiw_hdmiedid = VLV_HDMIW_HDMIEDID(pipe);
> >   		aud_config = VLV_AUD_CFG(pipe);
> >   		aud_cntl_st = VLV_AUD_CNTL_ST(pipe);
> > @@ -632,8 +632,8 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *crtc_state,
> >   			      const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	struct i915_audio_component *acomp = dev_priv->audio_component;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	struct i915_audio_component *acomp = i915->audio_component;
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct drm_connector *connector = conn_state->connector;
> >   	const struct drm_display_mode *adjusted_mode =
> > @@ -654,17 +654,17 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
> >   	connector->eld[6] = drm_av_sync_delay(connector, adjusted_mode) / 2;
> > -	if (dev_priv->display.audio_codec_enable)
> > -		dev_priv->display.audio_codec_enable(encoder,
> > +	if (i915->display.audio_codec_enable)
> > +		i915->display.audio_codec_enable(encoder,
> >   						     crtc_state,
> >   						     conn_state);
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> >   	encoder->audio_connector = connector;
> >   	/* referred in audio callbacks */
> > -	dev_priv->av_enc_map[pipe] = encoder;
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	i915->av_enc_map[pipe] = encoder;
> > +	mutex_unlock(&i915->av_mutex);
> >   	if (acomp && acomp->base.audio_ops &&
> >   	    acomp->base.audio_ops->pin_eld_notify) {
> > @@ -675,7 +675,7 @@ void intel_audio_codec_enable(struct intel_encoder *encoder,
> >   						 (int) port, (int) pipe);
> >   	}
> > -	intel_lpe_audio_notify(dev_priv, pipe, port, connector->eld,
> > +	intel_lpe_audio_notify(i915, pipe, port, connector->eld,
> >   			       crtc_state->port_clock,
> >   			       intel_crtc_has_dp_encoder(crtc_state));
> >   }
> > @@ -693,21 +693,21 @@ void intel_audio_codec_disable(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *old_crtc_state,
> >   			       const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	struct i915_audio_component *acomp = dev_priv->audio_component;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	struct i915_audio_component *acomp = i915->audio_component;
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	enum port port = encoder->port;
> >   	enum pipe pipe = crtc->pipe;
> > -	if (dev_priv->display.audio_codec_disable)
> > -		dev_priv->display.audio_codec_disable(encoder,
> > +	if (i915->display.audio_codec_disable)
> > +		i915->display.audio_codec_disable(encoder,
> >   						      old_crtc_state,
> >   						      old_conn_state);
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> >   	encoder->audio_connector = NULL;
> > -	dev_priv->av_enc_map[pipe] = NULL;
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	i915->av_enc_map[pipe] = NULL;
> > +	mutex_unlock(&i915->av_mutex);
> >   	if (acomp && acomp->base.audio_ops &&
> >   	    acomp->base.audio_ops->pin_eld_notify) {
> > @@ -718,31 +718,31 @@ void intel_audio_codec_disable(struct intel_encoder *encoder,
> >   						 (int) port, (int) pipe);
> >   	}
> > -	intel_lpe_audio_notify(dev_priv, pipe, port, NULL, 0, false);
> > +	intel_lpe_audio_notify(i915, pipe, port, NULL, 0, false);
> >   }
> >   /**
> >    * intel_init_audio_hooks - Set up chip specific audio hooks
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -void intel_init_audio_hooks(struct drm_i915_private *dev_priv)
> > +void intel_init_audio_hooks(struct drm_i915_private *i915)
> >   {
> > -	if (IS_G4X(dev_priv)) {
> > -		dev_priv->display.audio_codec_enable = g4x_audio_codec_enable;
> > -		dev_priv->display.audio_codec_disable = g4x_audio_codec_disable;
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		dev_priv->display.audio_codec_enable = ilk_audio_codec_enable;
> > -		dev_priv->display.audio_codec_disable = ilk_audio_codec_disable;
> > -	} else if (IS_HASWELL(dev_priv) || INTEL_GEN(dev_priv) >= 8) {
> > -		dev_priv->display.audio_codec_enable = hsw_audio_codec_enable;
> > -		dev_priv->display.audio_codec_disable = hsw_audio_codec_disable;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > -		dev_priv->display.audio_codec_enable = ilk_audio_codec_enable;
> > -		dev_priv->display.audio_codec_disable = ilk_audio_codec_disable;
> > +	if (IS_G4X(i915)) {
> > +		i915->display.audio_codec_enable = g4x_audio_codec_enable;
> > +		i915->display.audio_codec_disable = g4x_audio_codec_disable;
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		i915->display.audio_codec_enable = ilk_audio_codec_enable;
> > +		i915->display.audio_codec_disable = ilk_audio_codec_disable;
> > +	} else if (IS_HASWELL(i915) || INTEL_GEN(i915) >= 8) {
> > +		i915->display.audio_codec_enable = hsw_audio_codec_enable;
> > +		i915->display.audio_codec_disable = hsw_audio_codec_disable;
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> > +		i915->display.audio_codec_enable = ilk_audio_codec_enable;
> > +		i915->display.audio_codec_disable = ilk_audio_codec_disable;
> >   	}
> >   }
> > -static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> > +static void glk_force_audio_cdclk(struct drm_i915_private *i915,
> >   				  bool enable)
> >   {
> >   	struct drm_modeset_acquire_ctx ctx;
> > @@ -750,7 +750,7 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> >   	int ret;
> >   	drm_modeset_acquire_init(&ctx, 0);
> > -	state = drm_atomic_state_alloc(&dev_priv->drm);
> > +	state = drm_atomic_state_alloc(&i915->drm);
> >   	if (WARN_ON(!state))
> >   		return;
> > @@ -762,11 +762,11 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> >   		enable ? 2 * 96000 : 0;
> >   	/*
> > -	 * Protects dev_priv->cdclk.force_min_cdclk
> > +	 * Protects i915->cdclk.force_min_cdclk
> >   	 * Need to lock this here in case we have no active pipes
> >   	 * and thus wouldn't lock it during the commit otherwise.
> >   	 */
> > -	ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> > +	ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> >   			       &ctx);
> >   	if (!ret)
> >   		ret = drm_atomic_commit(state);
> > @@ -787,18 +787,18 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv,
> >   static unsigned long i915_audio_component_get_power(struct device *kdev)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> >   	intel_wakeref_t ret;
> >   	/* Catch potential impedance mismatches before they occur! */
> >   	BUILD_BUG_ON(sizeof(intel_wakeref_t) > sizeof(unsigned long));
> > -	ret = intel_display_power_get(dev_priv, POWER_DOMAIN_AUDIO);
> > +	ret = intel_display_power_get(i915, POWER_DOMAIN_AUDIO);
> >   	/* Force CDCLK to 2*BCLK as long as we need audio to be powered. */
> > -	if (dev_priv->audio_power_refcount++ == 0)
> > -		if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> > -			glk_force_audio_cdclk(dev_priv, true);
> > +	if (i915->audio_power_refcount++ == 0)
> > +		if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> > +			glk_force_audio_cdclk(i915, true);
> >   	return ret;
> >   }
> > @@ -806,24 +806,24 @@ static unsigned long i915_audio_component_get_power(struct device *kdev)
> >   static void i915_audio_component_put_power(struct device *kdev,
> >   					   unsigned long cookie)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> >   	/* Stop forcing CDCLK to 2*BCLK if no need for audio to be powered. */
> > -	if (--dev_priv->audio_power_refcount == 0)
> > -		if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> > -			glk_force_audio_cdclk(dev_priv, false);
> > +	if (--i915->audio_power_refcount == 0)
> > +		if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> > +			glk_force_audio_cdclk(i915, false);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_AUDIO, cookie);
> > +	intel_display_power_put(i915, POWER_DOMAIN_AUDIO, cookie);
> >   }
> >   static void i915_audio_component_codec_wake_override(struct device *kdev,
> >   						     bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> >   	unsigned long cookie;
> >   	u32 tmp;
> > -	if (!IS_GEN(dev_priv, 9))
> > +	if (!IS_GEN(i915, 9))
> >   		return;
> >   	cookie = i915_audio_component_get_power(kdev);
> > @@ -850,12 +850,12 @@ static void i915_audio_component_codec_wake_override(struct device *kdev,
> >   /* Get CDCLK in kHz  */
> >   static int i915_audio_component_get_cdclk_freq(struct device *kdev)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> > -	if (WARN_ON_ONCE(!HAS_DDI(dev_priv)))
> > +	if (WARN_ON_ONCE(!HAS_DDI(i915)))
> >   		return -ENODEV;
> > -	return dev_priv->cdclk.hw.cdclk;
> > +	return i915->cdclk.hw.cdclk;
> >   }
> >   /*
> > @@ -868,17 +868,17 @@ static int i915_audio_component_get_cdclk_freq(struct device *kdev)
> >    *   will get the right intel_encoder with port matched
> >    * Non-MST & (pipe < 0): get the right intel_encoder with port matched
> >    */
> > -static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> > +static struct intel_encoder *get_saved_enc(struct drm_i915_private *i915,
> >   					       int port, int pipe)
> >   {
> >   	struct intel_encoder *encoder;
> >   	/* MST */
> >   	if (pipe >= 0) {
> > -		if (WARN_ON(pipe >= ARRAY_SIZE(dev_priv->av_enc_map)))
> > +		if (WARN_ON(pipe >= ARRAY_SIZE(i915->av_enc_map)))
> >   			return NULL;
> > -		encoder = dev_priv->av_enc_map[pipe];
> > +		encoder = i915->av_enc_map[pipe];
> >   		/*
> >   		 * when bootup, audio driver may not know it is
> >   		 * MST or not. So it will poll all the port & pipe
> > @@ -893,8 +893,8 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> >   	if (pipe > 0)
> >   		return NULL;
> > -	for_each_pipe(dev_priv, pipe) {
> > -		encoder = dev_priv->av_enc_map[pipe];
> > +	for_each_pipe(i915, pipe) {
> > +		encoder = i915->av_enc_map[pipe];
> >   		if (encoder == NULL)
> >   			continue;
> > @@ -911,21 +911,21 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv,
> >   static int i915_audio_component_sync_audio_rate(struct device *kdev, int port,
> >   						int pipe, int rate)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > -	struct i915_audio_component *acomp = dev_priv->audio_component;
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> > +	struct i915_audio_component *acomp = i915->audio_component;
> >   	struct intel_encoder *encoder;
> >   	struct intel_crtc *crtc;
> >   	unsigned long cookie;
> >   	int err = 0;
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		return 0;
> >   	cookie = i915_audio_component_get_power(kdev);
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> >   	/* 1. get the pipe */
> > -	encoder = get_saved_enc(dev_priv, port, pipe);
> > +	encoder = get_saved_enc(i915, port, pipe);
> >   	if (!encoder || !encoder->base.crtc) {
> >   		DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port));
> >   		err = -ENODEV;
> > @@ -940,7 +940,7 @@ static int i915_audio_component_sync_audio_rate(struct device *kdev, int port,
> >   	hsw_audio_config_update(encoder, crtc->config);
> >    unlock:
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	mutex_unlock(&i915->av_mutex);
> >   	i915_audio_component_put_power(kdev, cookie);
> >   	return err;
> >   }
> > @@ -949,17 +949,17 @@ static int i915_audio_component_get_eld(struct device *kdev, int port,
> >   					int pipe, bool *enabled,
> >   					unsigned char *buf, int max_bytes)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(kdev);
> >   	struct intel_encoder *intel_encoder;
> >   	const u8 *eld;
> >   	int ret = -EINVAL;
> > -	mutex_lock(&dev_priv->av_mutex);
> > +	mutex_lock(&i915->av_mutex);
> > -	intel_encoder = get_saved_enc(dev_priv, port, pipe);
> > +	intel_encoder = get_saved_enc(i915, port, pipe);
> >   	if (!intel_encoder) {
> >   		DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port));
> > -		mutex_unlock(&dev_priv->av_mutex);
> > +		mutex_unlock(&i915->av_mutex);
> >   		return ret;
> >   	}
> > @@ -971,7 +971,7 @@ static int i915_audio_component_get_eld(struct device *kdev, int port,
> >   		memcpy(buf, eld, min(max_bytes, ret));
> >   	}
> > -	mutex_unlock(&dev_priv->av_mutex);
> > +	mutex_unlock(&i915->av_mutex);
> >   	return ret;
> >   }
> > @@ -989,7 +989,7 @@ static int i915_audio_component_bind(struct device *i915_kdev,
> >   				     struct device *hda_kdev, void *data)
> >   {
> >   	struct i915_audio_component *acomp = data;
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
> >   	int i;
> >   	if (WARN_ON(acomp->base.ops || acomp->base.dev))
> > @@ -998,14 +998,14 @@ static int i915_audio_component_bind(struct device *i915_kdev,
> >   	if (WARN_ON(!device_link_add(hda_kdev, i915_kdev, DL_FLAG_STATELESS)))
> >   		return -ENOMEM;
> > -	drm_modeset_lock_all(&dev_priv->drm);
> > +	drm_modeset_lock_all(&i915->drm);
> >   	acomp->base.ops = &i915_audio_component_ops;
> >   	acomp->base.dev = i915_kdev;
> >   	BUILD_BUG_ON(MAX_PORTS != I915_MAX_PORTS);
> >   	for (i = 0; i < ARRAY_SIZE(acomp->aud_sample_rate); i++)
> >   		acomp->aud_sample_rate[i] = 0;
> > -	dev_priv->audio_component = acomp;
> > -	drm_modeset_unlock_all(&dev_priv->drm);
> > +	i915->audio_component = acomp;
> > +	drm_modeset_unlock_all(&i915->drm);
> >   	return 0;
> >   }
> > @@ -1014,13 +1014,13 @@ static void i915_audio_component_unbind(struct device *i915_kdev,
> >   					struct device *hda_kdev, void *data)
> >   {
> >   	struct i915_audio_component *acomp = data;
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
> > -	drm_modeset_lock_all(&dev_priv->drm);
> > +	drm_modeset_lock_all(&i915->drm);
> >   	acomp->base.ops = NULL;
> >   	acomp->base.dev = NULL;
> > -	dev_priv->audio_component = NULL;
> > -	drm_modeset_unlock_all(&dev_priv->drm);
> > +	i915->audio_component = NULL;
> > +	drm_modeset_unlock_all(&i915->drm);
> >   	device_link_remove(hda_kdev, i915_kdev);
> >   }
> > @@ -1032,7 +1032,7 @@ static const struct component_ops i915_audio_component_bind_ops = {
> >   /**
> >    * i915_audio_component_init - initialize and register the audio component
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This will register with the component framework a child component which
> >    * will bind dynamically to the snd_hda_intel driver's corresponding master
> > @@ -1046,11 +1046,11 @@ static const struct component_ops i915_audio_component_bind_ops = {
> >    * We ignore any error during registration and continue with reduced
> >    * functionality (i.e. without HDMI audio).
> >    */
> > -static void i915_audio_component_init(struct drm_i915_private *dev_priv)
> > +static void i915_audio_component_init(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	ret = component_add_typed(dev_priv->drm.dev,
> > +	ret = component_add_typed(i915->drm.dev,
> >   				  &i915_audio_component_bind_ops,
> >   				  I915_COMPONENT_AUDIO);
> >   	if (ret < 0) {
> > @@ -1059,46 +1059,46 @@ static void i915_audio_component_init(struct drm_i915_private *dev_priv)
> >   		return;
> >   	}
> > -	dev_priv->audio_component_registered = true;
> > +	i915->audio_component_registered = true;
> >   }
> >   /**
> >    * i915_audio_component_cleanup - deregister the audio component
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Deregisters the audio component, breaking any existing binding to the
> >    * corresponding snd_hda_intel driver's master component.
> >    */
> > -static void i915_audio_component_cleanup(struct drm_i915_private *dev_priv)
> > +static void i915_audio_component_cleanup(struct drm_i915_private *i915)
> >   {
> > -	if (!dev_priv->audio_component_registered)
> > +	if (!i915->audio_component_registered)
> >   		return;
> > -	component_del(dev_priv->drm.dev, &i915_audio_component_bind_ops);
> > -	dev_priv->audio_component_registered = false;
> > +	component_del(i915->drm.dev, &i915_audio_component_bind_ops);
> > +	i915->audio_component_registered = false;
> >   }
> >   /**
> >    * intel_audio_init() - Initialize the audio driver either using
> >    * component framework or using lpe audio bridge
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    *
> >    */
> > -void intel_audio_init(struct drm_i915_private *dev_priv)
> > +void intel_audio_init(struct drm_i915_private *i915)
> >   {
> > -	if (intel_lpe_audio_init(dev_priv) < 0)
> > -		i915_audio_component_init(dev_priv);
> > +	if (intel_lpe_audio_init(i915) < 0)
> > +		i915_audio_component_init(i915);
> >   }
> >   /**
> >    * intel_audio_deinit() - deinitialize the audio driver
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    *
> >    */
> > -void intel_audio_deinit(struct drm_i915_private *dev_priv)
> > +void intel_audio_deinit(struct drm_i915_private *i915)
> >   {
> > -	if ((dev_priv)->lpe_audio.platdev != NULL)
> > -		intel_lpe_audio_teardown(dev_priv);
> > +	if ((i915)->lpe_audio.platdev != NULL)
> > +		intel_lpe_audio_teardown(i915);
> >   	else
> > -		i915_audio_component_cleanup(dev_priv);
> > +		i915_audio_component_cleanup(i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_audio.h b/drivers/gpu/drm/i915/intel_audio.h
> > index a3657c7a7ba2..c8fde42a9fde 100644
> > --- a/drivers/gpu/drm/i915/intel_audio.h
> > +++ b/drivers/gpu/drm/i915/intel_audio.h
> > @@ -11,14 +11,14 @@ struct drm_i915_private;
> >   struct intel_crtc_state;
> >   struct intel_encoder;
> > -void intel_init_audio_hooks(struct drm_i915_private *dev_priv);
> > +void intel_init_audio_hooks(struct drm_i915_private *i915);
> >   void intel_audio_codec_enable(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *crtc_state,
> >   			      const struct drm_connector_state *conn_state);
> >   void intel_audio_codec_disable(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *old_crtc_state,
> >   			       const struct drm_connector_state *old_conn_state);
> > -void intel_audio_init(struct drm_i915_private *dev_priv);
> > -void intel_audio_deinit(struct drm_i915_private *dev_priv);
> > +void intel_audio_init(struct drm_i915_private *i915);
> > +void intel_audio_deinit(struct drm_i915_private *i915);
> >   #endif /* __INTEL_AUDIO_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_bios.c b/drivers/gpu/drm/i915/intel_bios.c
> > index 1c037dfa83f5..7bdac55d08cc 100644
> > --- a/drivers/gpu/drm/i915/intel_bios.c
> > +++ b/drivers/gpu/drm/i915/intel_bios.c
> > @@ -202,7 +202,7 @@ get_lvds_fp_timing(const struct bdb_header *bdb,
> >   /* Try to find integrated panel data */
> >   static void
> > -parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> > +parse_lfp_panel_data(struct drm_i915_private *i915,
> >   		     const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_lvds_options *lvds_options;
> > @@ -219,9 +219,9 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> >   	if (!lvds_options)
> >   		return;
> > -	dev_priv->vbt.lvds_dither = lvds_options->pixel_dither;
> > +	i915->vbt.lvds_dither = lvds_options->pixel_dither;
> > -	ret = intel_opregion_get_panel_type(dev_priv);
> > +	ret = intel_opregion_get_panel_type(i915);
> >   	if (ret >= 0) {
> >   		WARN_ON(ret > 0xf);
> >   		panel_type = ret;
> > @@ -236,7 +236,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> >   		DRM_DEBUG_KMS("Panel type: %d (VBT)\n", panel_type);
> >   	}
> > -	dev_priv->vbt.panel_type = panel_type;
> > +	i915->vbt.panel_type = panel_type;
> >   	drrs_mode = (lvds_options->dps_panel_type_bits
> >   				>> (panel_type * 2)) & MODE_MASK;
> > @@ -247,15 +247,15 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> >   	 */
> >   	switch (drrs_mode) {
> >   	case 0:
> > -		dev_priv->vbt.drrs_type = STATIC_DRRS_SUPPORT;
> > +		i915->vbt.drrs_type = STATIC_DRRS_SUPPORT;
> >   		DRM_DEBUG_KMS("DRRS supported mode is static\n");
> >   		break;
> >   	case 2:
> > -		dev_priv->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT;
> > +		i915->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT;
> >   		DRM_DEBUG_KMS("DRRS supported mode is seamless\n");
> >   		break;
> >   	default:
> > -		dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> > +		i915->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> >   		DRM_DEBUG_KMS("DRRS not supported (VBT input)\n");
> >   		break;
> >   	}
> > @@ -278,7 +278,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> >   	fill_detail_timing_data(panel_fixed_mode, panel_dvo_timing);
> > -	dev_priv->vbt.lfp_lvds_vbt_mode = panel_fixed_mode;
> > +	i915->vbt.lfp_lvds_vbt_mode = panel_fixed_mode;
> >   	DRM_DEBUG_KMS("Found panel mode in BIOS VBT tables:\n");
> >   	drm_mode_debug_printmodeline(panel_fixed_mode);
> > @@ -290,20 +290,20 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv,
> >   		/* check the resolution, just to be sure */
> >   		if (fp_timing->x_res == panel_fixed_mode->hdisplay &&
> >   		    fp_timing->y_res == panel_fixed_mode->vdisplay) {
> > -			dev_priv->vbt.bios_lvds_val = fp_timing->lvds_reg_val;
> > +			i915->vbt.bios_lvds_val = fp_timing->lvds_reg_val;
> >   			DRM_DEBUG_KMS("VBT initial LVDS value %x\n",
> > -				      dev_priv->vbt.bios_lvds_val);
> > +				      i915->vbt.bios_lvds_val);
> >   		}
> >   	}
> >   }
> >   static void
> > -parse_lfp_backlight(struct drm_i915_private *dev_priv,
> > +parse_lfp_backlight(struct drm_i915_private *i915,
> >   		    const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_lfp_backlight_data *backlight_data;
> >   	const struct lfp_backlight_data_entry *entry;
> > -	int panel_type = dev_priv->vbt.panel_type;
> > +	int panel_type = i915->vbt.panel_type;
> >   	backlight_data = find_section(bdb, BDB_LVDS_BACKLIGHT);
> >   	if (!backlight_data)
> > @@ -317,38 +317,38 @@ parse_lfp_backlight(struct drm_i915_private *dev_priv,
> >   	entry = &backlight_data->data[panel_type];
> > -	dev_priv->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM;
> > -	if (!dev_priv->vbt.backlight.present) {
> > +	i915->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM;
> > +	if (!i915->vbt.backlight.present) {
> >   		DRM_DEBUG_KMS("PWM backlight not present in VBT (type %u)\n",
> >   			      entry->type);
> >   		return;
> >   	}
> > -	dev_priv->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI;
> > +	i915->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI;
> >   	if (bdb->version >= 191 &&
> >   	    get_blocksize(backlight_data) >= sizeof(*backlight_data)) {
> >   		const struct lfp_backlight_control_method *method;
> >   		method = &backlight_data->backlight_control[panel_type];
> > -		dev_priv->vbt.backlight.type = method->type;
> > -		dev_priv->vbt.backlight.controller = method->controller;
> > +		i915->vbt.backlight.type = method->type;
> > +		i915->vbt.backlight.controller = method->controller;
> >   	}
> > -	dev_priv->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz;
> > -	dev_priv->vbt.backlight.active_low_pwm = entry->active_low_pwm;
> > -	dev_priv->vbt.backlight.min_brightness = entry->min_brightness;
> > +	i915->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz;
> > +	i915->vbt.backlight.active_low_pwm = entry->active_low_pwm;
> > +	i915->vbt.backlight.min_brightness = entry->min_brightness;
> >   	DRM_DEBUG_KMS("VBT backlight PWM modulation frequency %u Hz, "
> >   		      "active %s, min brightness %u, level %u, controller %u\n",
> > -		      dev_priv->vbt.backlight.pwm_freq_hz,
> > -		      dev_priv->vbt.backlight.active_low_pwm ? "low" : "high",
> > -		      dev_priv->vbt.backlight.min_brightness,
> > +		      i915->vbt.backlight.pwm_freq_hz,
> > +		      i915->vbt.backlight.active_low_pwm ? "low" : "high",
> > +		      i915->vbt.backlight.min_brightness,
> >   		      backlight_data->level[panel_type],
> > -		      dev_priv->vbt.backlight.controller);
> > +		      i915->vbt.backlight.controller);
> >   }
> >   /* Try to find sdvo panel data */
> >   static void
> > -parse_sdvo_panel_data(struct drm_i915_private *dev_priv,
> > +parse_sdvo_panel_data(struct drm_i915_private *i915,
> >   		      const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_sdvo_panel_dtds *dtds;
> > @@ -381,16 +381,16 @@ parse_sdvo_panel_data(struct drm_i915_private *dev_priv,
> >   	fill_detail_timing_data(panel_fixed_mode, &dtds->dtds[index]);
> > -	dev_priv->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode;
> > +	i915->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode;
> >   	DRM_DEBUG_KMS("Found SDVO panel mode in BIOS VBT tables:\n");
> >   	drm_mode_debug_printmodeline(panel_fixed_mode);
> >   }
> > -static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv,
> > +static int intel_bios_ssc_frequency(struct drm_i915_private *i915,
> >   				    bool alternate)
> >   {
> > -	switch (INTEL_GEN(dev_priv)) {
> > +	switch (INTEL_GEN(i915)) {
> >   	case 2:
> >   		return alternate ? 66667 : 48000;
> >   	case 3:
> > @@ -402,7 +402,7 @@ static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -parse_general_features(struct drm_i915_private *dev_priv,
> > +parse_general_features(struct drm_i915_private *i915,
> >   		       const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_general_features *general;
> > @@ -411,30 +411,30 @@ parse_general_features(struct drm_i915_private *dev_priv,
> >   	if (!general)
> >   		return;
> > -	dev_priv->vbt.int_tv_support = general->int_tv_support;
> > +	i915->vbt.int_tv_support = general->int_tv_support;
> >   	/* int_crt_support can't be trusted on earlier platforms */
> >   	if (bdb->version >= 155 &&
> > -	    (HAS_DDI(dev_priv) || IS_VALLEYVIEW(dev_priv)))
> > -		dev_priv->vbt.int_crt_support = general->int_crt_support;
> > -	dev_priv->vbt.lvds_use_ssc = general->enable_ssc;
> > -	dev_priv->vbt.lvds_ssc_freq =
> > -		intel_bios_ssc_frequency(dev_priv, general->ssc_freq);
> > -	dev_priv->vbt.display_clock_mode = general->display_clock_mode;
> > -	dev_priv->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted;
> > +	    (HAS_DDI(i915) || IS_VALLEYVIEW(i915)))
> > +		i915->vbt.int_crt_support = general->int_crt_support;
> > +	i915->vbt.lvds_use_ssc = general->enable_ssc;
> > +	i915->vbt.lvds_ssc_freq =
> > +		intel_bios_ssc_frequency(i915, general->ssc_freq);
> > +	i915->vbt.display_clock_mode = general->display_clock_mode;
> > +	i915->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted;
> >   	if (bdb->version >= 181) {
> > -		dev_priv->vbt.orientation = general->rotate_180 ?
> > +		i915->vbt.orientation = general->rotate_180 ?
> >   			DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP :
> >   			DRM_MODE_PANEL_ORIENTATION_NORMAL;
> >   	} else {
> > -		dev_priv->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> > +		i915->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> >   	}
> >   	DRM_DEBUG_KMS("BDB_GENERAL_FEATURES int_tv_support %d int_crt_support %d lvds_use_ssc %d lvds_ssc_freq %d display_clock_mode %d fdi_rx_polarity_inverted %d\n",
> > -		      dev_priv->vbt.int_tv_support,
> > -		      dev_priv->vbt.int_crt_support,
> > -		      dev_priv->vbt.lvds_use_ssc,
> > -		      dev_priv->vbt.lvds_ssc_freq,
> > -		      dev_priv->vbt.display_clock_mode,
> > -		      dev_priv->vbt.fdi_rx_polarity_inverted);
> > +		      i915->vbt.int_tv_support,
> > +		      i915->vbt.int_crt_support,
> > +		      i915->vbt.lvds_use_ssc,
> > +		      i915->vbt.lvds_ssc_freq,
> > +		      i915->vbt.display_clock_mode,
> > +		      i915->vbt.fdi_rx_polarity_inverted);
> >   }
> >   static const struct child_device_config *
> > @@ -444,7 +444,7 @@ child_device_ptr(const struct bdb_general_definitions *defs, int i)
> >   }
> >   static void
> > -parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> > +parse_sdvo_device_mapping(struct drm_i915_private *i915, u8 bdb_version)
> >   {
> >   	struct sdvo_device_mapping *mapping;
> >   	const struct child_device_config *child;
> > @@ -454,13 +454,13 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> >   	 * Only parse SDVO mappings on gens that could have SDVO. This isn't
> >   	 * accurate and doesn't have to be, as long as it's not too strict.
> >   	 */
> > -	if (!IS_GEN_RANGE(dev_priv, 3, 7)) {
> > +	if (!IS_GEN_RANGE(i915, 3, 7)) {
> >   		DRM_DEBUG_KMS("Skipping SDVO device mapping\n");
> >   		return;
> >   	}
> > -	for (i = 0, count = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0, count = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		if (child->slave_addr != SLAVE_ADDR1 &&
> >   		    child->slave_addr != SLAVE_ADDR2) {
> > @@ -481,7 +481,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> >   			      child->slave_addr,
> >   			      (child->dvo_port == DEVICE_PORT_DVOB) ?
> >   			      "SDVOB" : "SDVOC");
> > -		mapping = &dev_priv->vbt.sdvo_mappings[child->dvo_port - 1];
> > +		mapping = &i915->vbt.sdvo_mappings[child->dvo_port - 1];
> >   		if (!mapping->initialized) {
> >   			mapping->dvo_port = child->dvo_port;
> >   			mapping->slave_addr = child->slave_addr;
> > @@ -515,7 +515,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version)
> >   }
> >   static void
> > -parse_driver_features(struct drm_i915_private *dev_priv,
> > +parse_driver_features(struct drm_i915_private *i915,
> >   		      const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_driver_features *driver;
> > @@ -524,14 +524,14 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> >   	if (!driver)
> >   		return;
> > -	if (INTEL_GEN(dev_priv) >= 5) {
> > +	if (INTEL_GEN(i915) >= 5) {
> >   		/*
> >   		 * Note that we consider BDB_DRIVER_FEATURE_INT_SDVO_LVDS
> >   		 * to mean "eDP". The VBT spec doesn't agree with that
> >   		 * interpretation, but real world VBTs seem to.
> >   		 */
> >   		if (driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS)
> > -			dev_priv->vbt.int_lvds_support = 0;
> > +			i915->vbt.int_lvds_support = 0;
> >   	} else {
> >   		/*
> >   		 * FIXME it's not clear which BDB version has the LVDS config
> > @@ -547,7 +547,7 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> >   		if (bdb->version >= 134 &&
> >   		    driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS &&
> >   		    driver->lvds_config != BDB_DRIVER_FEATURE_INT_SDVO_LVDS)
> > -			dev_priv->vbt.int_lvds_support = 0;
> > +			i915->vbt.int_lvds_support = 0;
> >   	}
> >   	DRM_DEBUG_KMS("DRRS State Enabled:%d\n", driver->drrs_enabled);
> > @@ -558,17 +558,17 @@ parse_driver_features(struct drm_i915_private *dev_priv,
> >   	 * driver->drrs_enabled=false
> >   	 */
> >   	if (!driver->drrs_enabled)
> > -		dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> > -	dev_priv->vbt.psr.enable = driver->psr_enabled;
> > +		i915->vbt.drrs_type = DRRS_NOT_SUPPORTED;
> > +	i915->vbt.psr.enable = driver->psr_enabled;
> >   }
> >   static void
> > -parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> > +parse_edp(struct drm_i915_private *i915, const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_edp *edp;
> >   	const struct edp_power_seq *edp_pps;
> >   	const struct edp_fast_link_params *edp_link_params;
> > -	int panel_type = dev_priv->vbt.panel_type;
> > +	int panel_type = i915->vbt.panel_type;
> >   	edp = find_section(bdb, BDB_EDP);
> >   	if (!edp)
> > @@ -576,13 +576,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	switch ((edp->color_depth >> (panel_type * 2)) & 3) {
> >   	case EDP_18BPP:
> > -		dev_priv->vbt.edp.bpp = 18;
> > +		i915->vbt.edp.bpp = 18;
> >   		break;
> >   	case EDP_24BPP:
> > -		dev_priv->vbt.edp.bpp = 24;
> > +		i915->vbt.edp.bpp = 24;
> >   		break;
> >   	case EDP_30BPP:
> > -		dev_priv->vbt.edp.bpp = 30;
> > +		i915->vbt.edp.bpp = 30;
> >   		break;
> >   	}
> > @@ -590,14 +590,14 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	edp_pps = &edp->power_seqs[panel_type];
> >   	edp_link_params = &edp->fast_link_params[panel_type];
> > -	dev_priv->vbt.edp.pps = *edp_pps;
> > +	i915->vbt.edp.pps = *edp_pps;
> >   	switch (edp_link_params->rate) {
> >   	case EDP_RATE_1_62:
> > -		dev_priv->vbt.edp.rate = DP_LINK_BW_1_62;
> > +		i915->vbt.edp.rate = DP_LINK_BW_1_62;
> >   		break;
> >   	case EDP_RATE_2_7:
> > -		dev_priv->vbt.edp.rate = DP_LINK_BW_2_7;
> > +		i915->vbt.edp.rate = DP_LINK_BW_2_7;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_KMS("VBT has unknown eDP link rate value %u\n",
> > @@ -607,13 +607,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	switch (edp_link_params->lanes) {
> >   	case EDP_LANE_1:
> > -		dev_priv->vbt.edp.lanes = 1;
> > +		i915->vbt.edp.lanes = 1;
> >   		break;
> >   	case EDP_LANE_2:
> > -		dev_priv->vbt.edp.lanes = 2;
> > +		i915->vbt.edp.lanes = 2;
> >   		break;
> >   	case EDP_LANE_4:
> > -		dev_priv->vbt.edp.lanes = 4;
> > +		i915->vbt.edp.lanes = 4;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_KMS("VBT has unknown eDP lane count value %u\n",
> > @@ -623,16 +623,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	switch (edp_link_params->preemphasis) {
> >   	case EDP_PREEMPHASIS_NONE:
> > -		dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0;
> > +		i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0;
> >   		break;
> >   	case EDP_PREEMPHASIS_3_5dB:
> > -		dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1;
> > +		i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1;
> >   		break;
> >   	case EDP_PREEMPHASIS_6dB:
> > -		dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2;
> > +		i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2;
> >   		break;
> >   	case EDP_PREEMPHASIS_9_5dB:
> > -		dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3;
> > +		i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_KMS("VBT has unknown eDP pre-emphasis value %u\n",
> > @@ -642,16 +642,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	switch (edp_link_params->vswing) {
> >   	case EDP_VSWING_0_4V:
> > -		dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0;
> > +		i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0;
> >   		break;
> >   	case EDP_VSWING_0_6V:
> > -		dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1;
> > +		i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1;
> >   		break;
> >   	case EDP_VSWING_0_8V:
> > -		dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> > +		i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> >   		break;
> >   	case EDP_VSWING_1_2V:
> > -		dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> > +		i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_KMS("VBT has unknown eDP voltage swing value %u\n",
> > @@ -664,21 +664,21 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   		/* Don't read from VBT if module parameter has valid value*/
> >   		if (i915_modparams.edp_vswing) {
> > -			dev_priv->vbt.edp.low_vswing =
> > +			i915->vbt.edp.low_vswing =
> >   				i915_modparams.edp_vswing == 1;
> >   		} else {
> >   			vswing = (edp->edp_vswing_preemph >> (panel_type * 4)) & 0xF;
> > -			dev_priv->vbt.edp.low_vswing = vswing == 0;
> > +			i915->vbt.edp.low_vswing = vswing == 0;
> >   		}
> >   	}
> >   }
> >   static void
> > -parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> > +parse_psr(struct drm_i915_private *i915, const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_psr *psr;
> >   	const struct psr_table *psr_table;
> > -	int panel_type = dev_priv->vbt.panel_type;
> > +	int panel_type = i915->vbt.panel_type;
> >   	psr = find_section(bdb, BDB_PSR);
> >   	if (!psr) {
> > @@ -688,25 +688,25 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	psr_table = &psr->psr_table[panel_type];
> > -	dev_priv->vbt.psr.full_link = psr_table->full_link;
> > -	dev_priv->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup;
> > +	i915->vbt.psr.full_link = psr_table->full_link;
> > +	i915->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup;
> >   	/* Allowed VBT values goes from 0 to 15 */
> > -	dev_priv->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 :
> > +	i915->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 :
> >   		psr_table->idle_frames > 15 ? 15 : psr_table->idle_frames;
> >   	switch (psr_table->lines_to_wait) {
> >   	case 0:
> > -		dev_priv->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT;
> > +		i915->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT;
> >   		break;
> >   	case 1:
> > -		dev_priv->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT;
> > +		i915->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT;
> >   		break;
> >   	case 2:
> > -		dev_priv->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT;
> > +		i915->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT;
> >   		break;
> >   	case 3:
> > -		dev_priv->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT;
> > +		i915->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_KMS("VBT has unknown PSR lines to wait %u\n",
> > @@ -719,48 +719,48 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   	 * Old decimal value is wake up time in multiples of 100 us.
> >   	 */
> >   	if (bdb->version >= 205 &&
> > -	    (IS_GEN9_BC(dev_priv) || IS_GEMINILAKE(dev_priv) ||
> > -	     INTEL_GEN(dev_priv) >= 10)) {
> > +	    (IS_GEN9_BC(i915) || IS_GEMINILAKE(i915) ||
> > +	     INTEL_GEN(i915) >= 10)) {
> >   		switch (psr_table->tp1_wakeup_time) {
> >   		case 0:
> > -			dev_priv->vbt.psr.tp1_wakeup_time_us = 500;
> > +			i915->vbt.psr.tp1_wakeup_time_us = 500;
> >   			break;
> >   		case 1:
> > -			dev_priv->vbt.psr.tp1_wakeup_time_us = 100;
> > +			i915->vbt.psr.tp1_wakeup_time_us = 100;
> >   			break;
> >   		case 3:
> > -			dev_priv->vbt.psr.tp1_wakeup_time_us = 0;
> > +			i915->vbt.psr.tp1_wakeup_time_us = 0;
> >   			break;
> >   		default:
> >   			DRM_DEBUG_KMS("VBT tp1 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n",
> >   					psr_table->tp1_wakeup_time);
> >   			/* fallthrough */
> >   		case 2:
> > -			dev_priv->vbt.psr.tp1_wakeup_time_us = 2500;
> > +			i915->vbt.psr.tp1_wakeup_time_us = 2500;
> >   			break;
> >   		}
> >   		switch (psr_table->tp2_tp3_wakeup_time) {
> >   		case 0:
> > -			dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 500;
> > +			i915->vbt.psr.tp2_tp3_wakeup_time_us = 500;
> >   			break;
> >   		case 1:
> > -			dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 100;
> > +			i915->vbt.psr.tp2_tp3_wakeup_time_us = 100;
> >   			break;
> >   		case 3:
> > -			dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 0;
> > +			i915->vbt.psr.tp2_tp3_wakeup_time_us = 0;
> >   			break;
> >   		default:
> >   			DRM_DEBUG_KMS("VBT tp2_tp3 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n",
> >   					psr_table->tp2_tp3_wakeup_time);
> >   			/* fallthrough */
> >   		case 2:
> > -			dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 2500;
> > +			i915->vbt.psr.tp2_tp3_wakeup_time_us = 2500;
> >   		break;
> >   		}
> >   	} else {
> > -		dev_priv->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100;
> > -		dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100;
> > +		i915->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100;
> > +		i915->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100;
> >   	}
> >   	if (bdb->version >= 226) {
> > @@ -782,74 +782,74 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb)
> >   			wakeup_time = 2500;
> >   			break;
> >   		}
> > -		dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time;
> > +		i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time;
> >   	} else {
> >   		/* Reusing PSR1 wakeup time for PSR2 in older VBTs */
> > -		dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = dev_priv->vbt.psr.tp2_tp3_wakeup_time_us;
> > +		i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = i915->vbt.psr.tp2_tp3_wakeup_time_us;
> >   	}
> >   }
> > -static void parse_dsi_backlight_ports(struct drm_i915_private *dev_priv,
> > +static void parse_dsi_backlight_ports(struct drm_i915_private *i915,
> >   				      u16 version, enum port port)
> >   {
> > -	if (!dev_priv->vbt.dsi.config->dual_link || version < 197) {
> > -		dev_priv->vbt.dsi.bl_ports = BIT(port);
> > -		if (dev_priv->vbt.dsi.config->cabc_supported)
> > -			dev_priv->vbt.dsi.cabc_ports = BIT(port);
> > +	if (!i915->vbt.dsi.config->dual_link || version < 197) {
> > +		i915->vbt.dsi.bl_ports = BIT(port);
> > +		if (i915->vbt.dsi.config->cabc_supported)
> > +			i915->vbt.dsi.cabc_ports = BIT(port);
> >   		return;
> >   	}
> > -	switch (dev_priv->vbt.dsi.config->dl_dcs_backlight_ports) {
> > +	switch (i915->vbt.dsi.config->dl_dcs_backlight_ports) {
> >   	case DL_DCS_PORT_A:
> > -		dev_priv->vbt.dsi.bl_ports = BIT(PORT_A);
> > +		i915->vbt.dsi.bl_ports = BIT(PORT_A);
> >   		break;
> >   	case DL_DCS_PORT_C:
> > -		dev_priv->vbt.dsi.bl_ports = BIT(PORT_C);
> > +		i915->vbt.dsi.bl_ports = BIT(PORT_C);
> >   		break;
> >   	default:
> >   	case DL_DCS_PORT_A_AND_C:
> > -		dev_priv->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C);
> > +		i915->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C);
> >   		break;
> >   	}
> > -	if (!dev_priv->vbt.dsi.config->cabc_supported)
> > +	if (!i915->vbt.dsi.config->cabc_supported)
> >   		return;
> > -	switch (dev_priv->vbt.dsi.config->dl_dcs_cabc_ports) {
> > +	switch (i915->vbt.dsi.config->dl_dcs_cabc_ports) {
> >   	case DL_DCS_PORT_A:
> > -		dev_priv->vbt.dsi.cabc_ports = BIT(PORT_A);
> > +		i915->vbt.dsi.cabc_ports = BIT(PORT_A);
> >   		break;
> >   	case DL_DCS_PORT_C:
> > -		dev_priv->vbt.dsi.cabc_ports = BIT(PORT_C);
> > +		i915->vbt.dsi.cabc_ports = BIT(PORT_C);
> >   		break;
> >   	default:
> >   	case DL_DCS_PORT_A_AND_C:
> > -		dev_priv->vbt.dsi.cabc_ports =
> > +		i915->vbt.dsi.cabc_ports =
> >   					BIT(PORT_A) | BIT(PORT_C);
> >   		break;
> >   	}
> >   }
> >   static void
> > -parse_mipi_config(struct drm_i915_private *dev_priv,
> > +parse_mipi_config(struct drm_i915_private *i915,
> >   		  const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_mipi_config *start;
> >   	const struct mipi_config *config;
> >   	const struct mipi_pps_data *pps;
> > -	int panel_type = dev_priv->vbt.panel_type;
> > +	int panel_type = i915->vbt.panel_type;
> >   	enum port port;
> >   	/* parse MIPI blocks only if LFP type is MIPI */
> > -	if (!intel_bios_is_dsi_present(dev_priv, &port))
> > +	if (!intel_bios_is_dsi_present(i915, &port))
> >   		return;
> >   	/* Initialize this to undefined indicating no generic MIPI support */
> > -	dev_priv->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID;
> > +	i915->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID;
> >   	/* Block #40 is already parsed and panel_fixed_mode is
> > -	 * stored in dev_priv->lfp_lvds_vbt_mode
> > +	 * stored in i915->lfp_lvds_vbt_mode
> >   	 * resuse this when needed
> >   	 */
> > @@ -873,17 +873,17 @@ parse_mipi_config(struct drm_i915_private *dev_priv,
> >   	pps = &start->pps[panel_type];
> >   	/* store as of now full data. Trim when we realise all is not needed */
> > -	dev_priv->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL);
> > -	if (!dev_priv->vbt.dsi.config)
> > +	i915->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL);
> > +	if (!i915->vbt.dsi.config)
> >   		return;
> > -	dev_priv->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL);
> > -	if (!dev_priv->vbt.dsi.pps) {
> > -		kfree(dev_priv->vbt.dsi.config);
> > +	i915->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL);
> > +	if (!i915->vbt.dsi.pps) {
> > +		kfree(i915->vbt.dsi.config);
> >   		return;
> >   	}
> > -	parse_dsi_backlight_ports(dev_priv, bdb->version, port);
> > +	parse_dsi_backlight_ports(i915, bdb->version, port);
> >   	/* FIXME is the 90 vs. 270 correct? */
> >   	switch (config->rotation) {
> > @@ -892,25 +892,25 @@ parse_mipi_config(struct drm_i915_private *dev_priv,
> >   		 * Most (all?) VBTs claim 0 degrees despite having
> >   		 * an upside down panel, thus we do not trust this.
> >   		 */
> > -		dev_priv->vbt.dsi.orientation =
> > +		i915->vbt.dsi.orientation =
> >   			DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> >   		break;
> >   	case ENABLE_ROTATION_90:
> > -		dev_priv->vbt.dsi.orientation =
> > +		i915->vbt.dsi.orientation =
> >   			DRM_MODE_PANEL_ORIENTATION_RIGHT_UP;
> >   		break;
> >   	case ENABLE_ROTATION_180:
> > -		dev_priv->vbt.dsi.orientation =
> > +		i915->vbt.dsi.orientation =
> >   			DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP;
> >   		break;
> >   	case ENABLE_ROTATION_270:
> > -		dev_priv->vbt.dsi.orientation =
> > +		i915->vbt.dsi.orientation =
> >   			DRM_MODE_PANEL_ORIENTATION_LEFT_UP;
> >   		break;
> >   	}
> >   	/* We have mandatory mipi config blocks. Initialize as generic panel */
> > -	dev_priv->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID;
> > +	i915->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID;
> >   }
> >   /* Find the sequence block and size for the given panel. */
> > @@ -1073,12 +1073,12 @@ static int goto_next_sequence_v3(const u8 *data, int index, int total)
> >    * Get len of pre-fixed deassert fragment from a v1 init OTP sequence,
> >    * skip all delay + gpio operands and stop at the first DSI packet op.
> >    */
> > -static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv)
> > +static int get_init_otp_deassert_fragment_len(struct drm_i915_private *i915)
> >   {
> > -	const u8 *data = dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> > +	const u8 *data = i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> >   	int index, len;
> > -	if (WARN_ON(!data || dev_priv->vbt.dsi.seq_version != 1))
> > +	if (WARN_ON(!data || i915->vbt.dsi.seq_version != 1))
> >   		return 0;
> >   	/* index = 1 to skip sequence byte */
> > @@ -1106,54 +1106,54 @@ static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv)
> >    * these devices we split the init OTP sequence into a deassert sequence and
> >    * the actual init OTP part.
> >    */
> > -static void fixup_mipi_sequences(struct drm_i915_private *dev_priv)
> > +static void fixup_mipi_sequences(struct drm_i915_private *i915)
> >   {
> >   	u8 *init_otp;
> >   	int len;
> >   	/* Limit this to VLV for now. */
> > -	if (!IS_VALLEYVIEW(dev_priv))
> > +	if (!IS_VALLEYVIEW(i915))
> >   		return;
> >   	/* Limit this to v1 vid-mode sequences */
> > -	if (dev_priv->vbt.dsi.config->is_cmd_mode ||
> > -	    dev_priv->vbt.dsi.seq_version != 1)
> > +	if (i915->vbt.dsi.config->is_cmd_mode ||
> > +	    i915->vbt.dsi.seq_version != 1)
> >   		return;
> >   	/* Only do this if there are otp and assert seqs and no deassert seq */
> > -	if (!dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] ||
> > -	    !dev_priv->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] ||
> > -	    dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET])
> > +	if (!i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] ||
> > +	    !i915->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] ||
> > +	    i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET])
> >   		return;
> >   	/* The deassert-sequence ends at the first DSI packet */
> > -	len = get_init_otp_deassert_fragment_len(dev_priv);
> > +	len = get_init_otp_deassert_fragment_len(i915);
> >   	if (!len)
> >   		return;
> >   	DRM_DEBUG_KMS("Using init OTP fragment to deassert reset\n");
> >   	/* Copy the fragment, update seq byte and terminate it */
> > -	init_otp = (u8 *)dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> > -	dev_priv->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL);
> > -	if (!dev_priv->vbt.dsi.deassert_seq)
> > +	init_otp = (u8 *)i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP];
> > +	i915->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL);
> > +	if (!i915->vbt.dsi.deassert_seq)
> >   		return;
> > -	dev_priv->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET;
> > -	dev_priv->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END;
> > +	i915->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET;
> > +	i915->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END;
> >   	/* Use the copy for deassert */
> > -	dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] =
> > -		dev_priv->vbt.dsi.deassert_seq;
> > +	i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] =
> > +		i915->vbt.dsi.deassert_seq;
> >   	/* Replace the last byte of the fragment with init OTP seq byte */
> >   	init_otp[len - 1] = MIPI_SEQ_INIT_OTP;
> >   	/* And make MIPI_MIPI_SEQ_INIT_OTP point to it */
> > -	dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1;
> > +	i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1;
> >   }
> >   static void
> > -parse_mipi_sequence(struct drm_i915_private *dev_priv,
> > +parse_mipi_sequence(struct drm_i915_private *i915,
> >   		    const struct bdb_header *bdb)
> >   {
> > -	int panel_type = dev_priv->vbt.panel_type;
> > +	int panel_type = i915->vbt.panel_type;
> >   	const struct bdb_mipi_sequence *sequence;
> >   	const u8 *seq_data;
> >   	u32 seq_size;
> > @@ -1161,7 +1161,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> >   	int index = 0;
> >   	/* Only our generic panel driver uses the sequence block. */
> > -	if (dev_priv->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID)
> > +	if (i915->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID)
> >   		return;
> >   	sequence = find_section(bdb, BDB_MIPI_SEQUENCE);
> > @@ -1202,7 +1202,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> >   		if (seq_id == MIPI_SEQ_TEAR_ON || seq_id == MIPI_SEQ_TEAR_OFF)
> >   			DRM_DEBUG_KMS("Unsupported sequence %u\n", seq_id);
> > -		dev_priv->vbt.dsi.sequence[seq_id] = data + index;
> > +		i915->vbt.dsi.sequence[seq_id] = data + index;
> >   		if (sequence->version >= 3)
> >   			index = goto_next_sequence_v3(data, index, seq_size);
> > @@ -1214,18 +1214,18 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv,
> >   		}
> >   	}
> > -	dev_priv->vbt.dsi.data = data;
> > -	dev_priv->vbt.dsi.size = seq_size;
> > -	dev_priv->vbt.dsi.seq_version = sequence->version;
> > +	i915->vbt.dsi.data = data;
> > +	i915->vbt.dsi.size = seq_size;
> > +	i915->vbt.dsi.seq_version = sequence->version;
> > -	fixup_mipi_sequences(dev_priv);
> > +	fixup_mipi_sequences(i915);
> >   	DRM_DEBUG_DRIVER("MIPI related VBT parsing complete\n");
> >   	return;
> >   err:
> >   	kfree(data);
> > -	memset(dev_priv->vbt.dsi.sequence, 0, sizeof(dev_priv->vbt.dsi.sequence));
> > +	memset(i915->vbt.dsi.sequence, 0, sizeof(i915->vbt.dsi.sequence));
> >   }
> >   static u8 translate_iboost(u8 val)
> > @@ -1254,16 +1254,16 @@ static enum port get_port_by_ddc_pin(struct drm_i915_private *i915, u8 ddc_pin)
> >   	return PORT_NONE;
> >   }
> > -static void sanitize_ddc_pin(struct drm_i915_private *dev_priv,
> > +static void sanitize_ddc_pin(struct drm_i915_private *i915,
> >   			     enum port port)
> >   {
> > -	struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port];
> > +	struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port];
> >   	enum port p;
> >   	if (!info->alternate_ddc_pin)
> >   		return;
> > -	p = get_port_by_ddc_pin(dev_priv, info->alternate_ddc_pin);
> > +	p = get_port_by_ddc_pin(i915, info->alternate_ddc_pin);
> >   	if (p != PORT_NONE) {
> >   		DRM_DEBUG_KMS("port %c trying to use the same DDC pin (0x%x) as port %c, "
> >   			      "disabling port %c DVI/HDMI support\n",
> > @@ -1300,16 +1300,16 @@ static enum port get_port_by_aux_ch(struct drm_i915_private *i915, u8 aux_ch)
> >   	return PORT_NONE;
> >   }
> > -static void sanitize_aux_ch(struct drm_i915_private *dev_priv,
> > +static void sanitize_aux_ch(struct drm_i915_private *i915,
> >   			    enum port port)
> >   {
> > -	struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port];
> > +	struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port];
> >   	enum port p;
> >   	if (!info->alternate_aux_channel)
> >   		return;
> > -	p = get_port_by_aux_ch(dev_priv, info->alternate_aux_channel);
> > +	p = get_port_by_aux_ch(i915, info->alternate_aux_channel);
> >   	if (p != PORT_NONE) {
> >   		DRM_DEBUG_KMS("port %c trying to use the same AUX CH (0x%x) as port %c, "
> >   			      "disabling port %c DP support\n",
> > @@ -1347,15 +1347,15 @@ static const u8 icp_ddc_pin_map[] = {
> >   	[ICL_DDC_BUS_PORT_4] = GMBUS_PIN_12_TC4_ICP,
> >   };
> > -static u8 map_ddc_pin(struct drm_i915_private *dev_priv, u8 vbt_pin)
> > +static u8 map_ddc_pin(struct drm_i915_private *i915, u8 vbt_pin)
> >   {
> >   	const u8 *ddc_pin_map;
> >   	int n_entries;
> > -	if (HAS_PCH_ICP(dev_priv)) {
> > +	if (HAS_PCH_ICP(i915)) {
> >   		ddc_pin_map = icp_ddc_pin_map;
> >   		n_entries = ARRAY_SIZE(icp_ddc_pin_map);
> > -	} else if (HAS_PCH_CNP(dev_priv)) {
> > +	} else if (HAS_PCH_CNP(i915)) {
> >   		ddc_pin_map = cnp_ddc_pin_map;
> >   		n_entries = ARRAY_SIZE(cnp_ddc_pin_map);
> >   	} else {
> > @@ -1401,7 +1401,7 @@ static enum port dvo_port_to_port(u8 dvo_port)
> >   	return PORT_NONE;
> >   }
> > -static void parse_ddi_port(struct drm_i915_private *dev_priv,
> > +static void parse_ddi_port(struct drm_i915_private *i915,
> >   			   const struct child_device_config *child,
> >   			   u8 bdb_version)
> >   {
> > @@ -1413,7 +1413,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> >   	if (port == PORT_NONE)
> >   		return;
> > -	info = &dev_priv->vbt.ddi_port_info[port];
> > +	info = &i915->vbt.ddi_port_info[port];
> >   	if (info->child) {
> >   		DRM_DEBUG_KMS("More than one child device for port %c in VBT, using the first.\n",
> > @@ -1447,7 +1447,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> >   	DRM_DEBUG_KMS("Port %c VBT info: CRT:%d DVI:%d HDMI:%d DP:%d eDP:%d LSPCON:%d USB-Type-C:%d TBT:%d\n",
> >   		      port_name(port), is_crt, is_dvi, is_hdmi, is_dp, is_edp,
> > -		      HAS_LSPCON(dev_priv) && child->lspcon,
> > +		      HAS_LSPCON(i915) && child->lspcon,
> >   		      info->supports_typec_usb, info->supports_tbt);
> >   	if (is_edp && is_dvi)
> > @@ -1469,10 +1469,10 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> >   	if (is_dvi) {
> >   		u8 ddc_pin;
> > -		ddc_pin = map_ddc_pin(dev_priv, child->ddc_pin);
> > -		if (intel_gmbus_is_valid_pin(dev_priv, ddc_pin)) {
> > +		ddc_pin = map_ddc_pin(i915, child->ddc_pin);
> > +		if (intel_gmbus_is_valid_pin(i915, ddc_pin)) {
> >   			info->alternate_ddc_pin = ddc_pin;
> > -			sanitize_ddc_pin(dev_priv, port);
> > +			sanitize_ddc_pin(i915, port);
> >   		} else {
> >   			DRM_DEBUG_KMS("Port %c has invalid DDC pin %d, "
> >   				      "sticking to defaults\n",
> > @@ -1483,7 +1483,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> >   	if (is_dp) {
> >   		info->alternate_aux_channel = child->aux_channel;
> > -		sanitize_aux_ch(dev_priv, port);
> > +		sanitize_aux_ch(i915, port);
> >   	}
> >   	if (bdb_version >= 158) {
> > @@ -1553,26 +1553,26 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv,
> >   	info->child = child;
> >   }
> > -static void parse_ddi_ports(struct drm_i915_private *dev_priv, u8 bdb_version)
> > +static void parse_ddi_ports(struct drm_i915_private *i915, u8 bdb_version)
> >   {
> >   	const struct child_device_config *child;
> >   	int i;
> > -	if (!HAS_DDI(dev_priv) && !IS_CHERRYVIEW(dev_priv))
> > +	if (!HAS_DDI(i915) && !IS_CHERRYVIEW(i915))
> >   		return;
> >   	if (bdb_version < 155)
> >   		return;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> > -		parse_ddi_port(dev_priv, child, bdb_version);
> > +		parse_ddi_port(i915, child, bdb_version);
> >   	}
> >   }
> >   static void
> > -parse_general_definitions(struct drm_i915_private *dev_priv,
> > +parse_general_definitions(struct drm_i915_private *i915,
> >   			  const struct bdb_header *bdb)
> >   {
> >   	const struct bdb_general_definitions *defs;
> > @@ -1597,8 +1597,8 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> >   	bus_pin = defs->crt_ddc_gmbus_pin;
> >   	DRM_DEBUG_KMS("crt_ddc_bus_pin: %d\n", bus_pin);
> > -	if (intel_gmbus_is_valid_pin(dev_priv, bus_pin))
> > -		dev_priv->vbt.crt_ddc_pin = bus_pin;
> > +	if (intel_gmbus_is_valid_pin(i915, bus_pin))
> > +		i915->vbt.crt_ddc_pin = bus_pin;
> >   	if (bdb->version < 106) {
> >   		expected_size = 22;
> > @@ -1645,13 +1645,13 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> >   		DRM_DEBUG_KMS("no child dev is parsed from VBT\n");
> >   		return;
> >   	}
> > -	dev_priv->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL);
> > -	if (!dev_priv->vbt.child_dev) {
> > +	i915->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL);
> > +	if (!i915->vbt.child_dev) {
> >   		DRM_DEBUG_KMS("No memory space for child device\n");
> >   		return;
> >   	}
> > -	dev_priv->vbt.child_dev_num = count;
> > +	i915->vbt.child_dev_num = count;
> >   	count = 0;
> >   	for (i = 0; i < child_device_num; i++) {
> >   		child = child_device_ptr(defs, i);
> > @@ -1663,7 +1663,7 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> >   		 * (child_dev_size) of the child device. Accessing the data must
> >   		 * depend on VBT version.
> >   		 */
> > -		memcpy(dev_priv->vbt.child_dev + count, child,
> > +		memcpy(i915->vbt.child_dev + count, child,
> >   		       min_t(size_t, defs->child_dev_size, sizeof(*child)));
> >   		count++;
> >   	}
> > @@ -1671,41 +1671,41 @@ parse_general_definitions(struct drm_i915_private *dev_priv,
> >   /* Common defaults which may be overridden by VBT. */
> >   static void
> > -init_vbt_defaults(struct drm_i915_private *dev_priv)
> > +init_vbt_defaults(struct drm_i915_private *i915)
> >   {
> >   	enum port port;
> > -	dev_priv->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC;
> > +	i915->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC;
> >   	/* Default to having backlight */
> > -	dev_priv->vbt.backlight.present = true;
> > +	i915->vbt.backlight.present = true;
> >   	/* LFP panel data */
> > -	dev_priv->vbt.lvds_dither = 1;
> > +	i915->vbt.lvds_dither = 1;
> >   	/* SDVO panel data */
> > -	dev_priv->vbt.sdvo_lvds_vbt_mode = NULL;
> > +	i915->vbt.sdvo_lvds_vbt_mode = NULL;
> >   	/* general features */
> > -	dev_priv->vbt.int_tv_support = 1;
> > -	dev_priv->vbt.int_crt_support = 1;
> > +	i915->vbt.int_tv_support = 1;
> > +	i915->vbt.int_crt_support = 1;
> >   	/* driver features */
> > -	dev_priv->vbt.int_lvds_support = 1;
> > +	i915->vbt.int_lvds_support = 1;
> >   	/* Default to using SSC */
> > -	dev_priv->vbt.lvds_use_ssc = 1;
> > +	i915->vbt.lvds_use_ssc = 1;
> >   	/*
> >   	 * Core/SandyBridge/IvyBridge use alternative (120MHz) reference
> >   	 * clock for LVDS.
> >   	 */
> > -	dev_priv->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(dev_priv,
> > -			!HAS_PCH_SPLIT(dev_priv));
> > -	DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", dev_priv->vbt.lvds_ssc_freq);
> > +	i915->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(i915,
> > +			!HAS_PCH_SPLIT(i915));
> > +	DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", i915->vbt.lvds_ssc_freq);
> >   	for (port = PORT_A; port < I915_MAX_PORTS; port++) {
> >   		struct ddi_vbt_port_info *info =
> > -			&dev_priv->vbt.ddi_port_info[port];
> > +			&i915->vbt.ddi_port_info[port];
> >   		info->hdmi_level_shift = HDMI_LEVEL_SHIFT_UNKNOWN;
> >   	}
> > @@ -1713,19 +1713,19 @@ init_vbt_defaults(struct drm_i915_private *dev_priv)
> >   /* Defaults to initialize only if there is no VBT. */
> >   static void
> > -init_vbt_missing_defaults(struct drm_i915_private *dev_priv)
> > +init_vbt_missing_defaults(struct drm_i915_private *i915)
> >   {
> >   	enum port port;
> >   	for (port = PORT_A; port < I915_MAX_PORTS; port++) {
> >   		struct ddi_vbt_port_info *info =
> > -			&dev_priv->vbt.ddi_port_info[port];
> > +			&i915->vbt.ddi_port_info[port];
> >   		/*
> >   		 * VBT has the TypeC mode (native,TBT/USB) and we don't want
> >   		 * to detect it.
> >   		 */
> > -		if (intel_port_is_tc(dev_priv, port))
> > +		if (intel_port_is_tc(i915, port))
> >   			continue;
> >   		info->supports_dvi = (port != PORT_A && port != PORT_E);
> > @@ -1811,25 +1811,25 @@ static const struct vbt_header *find_vbt(void __iomem *bios, size_t size)
> >   /**
> >    * intel_bios_init - find VBT and initialize settings from the BIOS
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Parse and initialize settings from the Video BIOS Tables (VBT). If the VBT
> >    * was not found in ACPI OpRegion, try to find it in PCI ROM first. Also
> >    * initialize some defaults if the VBT is not present at all.
> >    */
> > -void intel_bios_init(struct drm_i915_private *dev_priv)
> > +void intel_bios_init(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > -	const struct vbt_header *vbt = dev_priv->opregion.vbt;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > +	const struct vbt_header *vbt = i915->opregion.vbt;
> >   	const struct bdb_header *bdb;
> >   	u8 __iomem *bios = NULL;
> > -	if (!HAS_DISPLAY(dev_priv)) {
> > +	if (!HAS_DISPLAY(i915)) {
> >   		DRM_DEBUG_KMS("Skipping VBT init due to disabled display.\n");
> >   		return;
> >   	}
> > -	init_vbt_defaults(dev_priv);
> > +	init_vbt_defaults(i915);
> >   	/* If the OpRegion does not have VBT, look in PCI ROM. */
> >   	if (!vbt) {
> > @@ -1852,25 +1852,25 @@ void intel_bios_init(struct drm_i915_private *dev_priv)
> >   		      (int)sizeof(vbt->signature), vbt->signature, bdb->version);
> >   	/* Grab useful general definitions */
> > -	parse_general_features(dev_priv, bdb);
> > -	parse_general_definitions(dev_priv, bdb);
> > -	parse_lfp_panel_data(dev_priv, bdb);
> > -	parse_lfp_backlight(dev_priv, bdb);
> > -	parse_sdvo_panel_data(dev_priv, bdb);
> > -	parse_driver_features(dev_priv, bdb);
> > -	parse_edp(dev_priv, bdb);
> > -	parse_psr(dev_priv, bdb);
> > -	parse_mipi_config(dev_priv, bdb);
> > -	parse_mipi_sequence(dev_priv, bdb);
> > +	parse_general_features(i915, bdb);
> > +	parse_general_definitions(i915, bdb);
> > +	parse_lfp_panel_data(i915, bdb);
> > +	parse_lfp_backlight(i915, bdb);
> > +	parse_sdvo_panel_data(i915, bdb);
> > +	parse_driver_features(i915, bdb);
> > +	parse_edp(i915, bdb);
> > +	parse_psr(i915, bdb);
> > +	parse_mipi_config(i915, bdb);
> > +	parse_mipi_sequence(i915, bdb);
> >   	/* Further processing on pre-parsed data */
> > -	parse_sdvo_device_mapping(dev_priv, bdb->version);
> > -	parse_ddi_ports(dev_priv, bdb->version);
> > +	parse_sdvo_device_mapping(i915, bdb->version);
> > +	parse_ddi_ports(i915, bdb->version);
> >   out:
> >   	if (!vbt) {
> >   		DRM_INFO("Failed to find VBIOS tables (VBT)\n");
> > -		init_vbt_missing_defaults(dev_priv);
> > +		init_vbt_missing_defaults(i915);
> >   	}
> >   	if (bios)
> > @@ -1879,47 +1879,47 @@ void intel_bios_init(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_bios_cleanup - Free any resources allocated by intel_bios_init()
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    */
> > -void intel_bios_cleanup(struct drm_i915_private *dev_priv)
> > +void intel_bios_cleanup(struct drm_i915_private *i915)
> >   {
> > -	kfree(dev_priv->vbt.child_dev);
> > -	dev_priv->vbt.child_dev = NULL;
> > -	dev_priv->vbt.child_dev_num = 0;
> > -	kfree(dev_priv->vbt.sdvo_lvds_vbt_mode);
> > -	dev_priv->vbt.sdvo_lvds_vbt_mode = NULL;
> > -	kfree(dev_priv->vbt.lfp_lvds_vbt_mode);
> > -	dev_priv->vbt.lfp_lvds_vbt_mode = NULL;
> > -	kfree(dev_priv->vbt.dsi.data);
> > -	dev_priv->vbt.dsi.data = NULL;
> > -	kfree(dev_priv->vbt.dsi.pps);
> > -	dev_priv->vbt.dsi.pps = NULL;
> > -	kfree(dev_priv->vbt.dsi.config);
> > -	dev_priv->vbt.dsi.config = NULL;
> > -	kfree(dev_priv->vbt.dsi.deassert_seq);
> > -	dev_priv->vbt.dsi.deassert_seq = NULL;
> > +	kfree(i915->vbt.child_dev);
> > +	i915->vbt.child_dev = NULL;
> > +	i915->vbt.child_dev_num = 0;
> > +	kfree(i915->vbt.sdvo_lvds_vbt_mode);
> > +	i915->vbt.sdvo_lvds_vbt_mode = NULL;
> > +	kfree(i915->vbt.lfp_lvds_vbt_mode);
> > +	i915->vbt.lfp_lvds_vbt_mode = NULL;
> > +	kfree(i915->vbt.dsi.data);
> > +	i915->vbt.dsi.data = NULL;
> > +	kfree(i915->vbt.dsi.pps);
> > +	i915->vbt.dsi.pps = NULL;
> > +	kfree(i915->vbt.dsi.config);
> > +	i915->vbt.dsi.config = NULL;
> > +	kfree(i915->vbt.dsi.deassert_seq);
> > +	i915->vbt.dsi.deassert_seq = NULL;
> >   }
> >   /**
> >    * intel_bios_is_tv_present - is integrated TV present in VBT
> > - * @dev_priv:	i915 device instance
> > + * @i915:	i915 device instance
> >    *
> >    * Return true if TV is present. If no child devices were parsed from VBT,
> >    * assume TV is present.
> >    */
> > -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv)
> > +bool intel_bios_is_tv_present(struct drm_i915_private *i915)
> >   {
> >   	const struct child_device_config *child;
> >   	int i;
> > -	if (!dev_priv->vbt.int_tv_support)
> > +	if (!i915->vbt.int_tv_support)
> >   		return false;
> > -	if (!dev_priv->vbt.child_dev_num)
> > +	if (!i915->vbt.child_dev_num)
> >   		return true;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		/*
> >   		 * If the device type is not TV, continue.
> >   		 */
> > @@ -1943,22 +1943,22 @@ bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_bios_is_lvds_present - is LVDS present in VBT
> > - * @dev_priv:	i915 device instance
> > + * @i915:	i915 device instance
> >    * @i2c_pin:	i2c pin for LVDS if present
> >    *
> >    * Return true if LVDS is present. If no child devices were parsed from VBT,
> >    * assume LVDS is present.
> >    */
> > -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> > +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin)
> >   {
> >   	const struct child_device_config *child;
> >   	int i;
> > -	if (!dev_priv->vbt.child_dev_num)
> > +	if (!i915->vbt.child_dev_num)
> >   		return true;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		/* If the device type is not LFP, continue.
> >   		 * We have to check both the new identifiers as well as the
> > @@ -1968,7 +1968,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> >   		    child->device_type != DEVICE_TYPE_LFP)
> >   			continue;
> > -		if (intel_gmbus_is_valid_pin(dev_priv, child->i2c_pin))
> > +		if (intel_gmbus_is_valid_pin(i915, child->i2c_pin))
> >   			*i2c_pin = child->i2c_pin;
> >   		/* However, we cannot trust the BIOS writers to populate
> > @@ -1984,7 +1984,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> >   		 * additional data.  Trust that if the VBT was written into
> >   		 * the OpRegion then they have validated the LVDS's existence.
> >   		 */
> > -		if (dev_priv->opregion.vbt)
> > +		if (i915->opregion.vbt)
> >   			return true;
> >   	}
> > @@ -1993,12 +1993,12 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin)
> >   /**
> >    * intel_bios_is_port_present - is the specified digital port present
> > - * @dev_priv:	i915 device instance
> > + * @i915:	i915 device instance
> >    * @port:	port to check
> >    *
> >    * Return true if the device in %port is present.
> >    */
> > -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port)
> > +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port)
> >   {
> >   	const struct child_device_config *child;
> >   	static const struct {
> > @@ -2012,9 +2012,9 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
> >   	};
> >   	int i;
> > -	if (HAS_DDI(dev_priv)) {
> > +	if (HAS_DDI(i915)) {
> >   		const struct ddi_vbt_port_info *port_info =
> > -			&dev_priv->vbt.ddi_port_info[port];
> > +			&i915->vbt.ddi_port_info[port];
> >   		return port_info->supports_dp ||
> >   		       port_info->supports_dvi ||
> > @@ -2025,11 +2025,11 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
> >   	if (WARN_ON(port == PORT_A) || port >= ARRAY_SIZE(port_mapping))
> >   		return false;
> > -	if (!dev_priv->vbt.child_dev_num)
> > +	if (!i915->vbt.child_dev_num)
> >   		return false;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		if ((child->dvo_port == port_mapping[port].dp ||
> >   		     child->dvo_port == port_mapping[port].hdmi) &&
> > @@ -2043,12 +2043,12 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por
> >   /**
> >    * intel_bios_is_port_edp - is the device in given port eDP
> > - * @dev_priv:	i915 device instance
> > + * @i915:	i915 device instance
> >    * @port:	port to check
> >    *
> >    * Return true if the device in %port is eDP.
> >    */
> > -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> > +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port)
> >   {
> >   	const struct child_device_config *child;
> >   	static const short port_mapping[] = {
> > @@ -2060,14 +2060,14 @@ bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> >   	};
> >   	int i;
> > -	if (HAS_DDI(dev_priv))
> > -		return dev_priv->vbt.ddi_port_info[port].supports_edp;
> > +	if (HAS_DDI(i915))
> > +		return i915->vbt.ddi_port_info[port].supports_edp;
> > -	if (!dev_priv->vbt.child_dev_num)
> > +	if (!i915->vbt.child_dev_num)
> >   		return false;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		if (child->dvo_port == port_mapping[port] &&
> >   		    (child->device_type & DEVICE_TYPE_eDP_BITS) ==
> > @@ -2113,14 +2113,14 @@ static bool child_dev_is_dp_dual_mode(const struct child_device_config *child,
> >   	return false;
> >   }
> > -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv,
> > +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915,
> >   				     enum port port)
> >   {
> >   	const struct child_device_config *child;
> >   	int i;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		if (child_dev_is_dp_dual_mode(child, port))
> >   			return true;
> > @@ -2131,20 +2131,20 @@ bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_bios_is_dsi_present - is DSI present in VBT
> > - * @dev_priv:	i915 device instance
> > + * @i915:	i915 device instance
> >    * @port:	port for DSI if present
> >    *
> >    * Return true if DSI is present, and return the port in %port.
> >    */
> > -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv,
> > +bool intel_bios_is_dsi_present(struct drm_i915_private *i915,
> >   			       enum port *port)
> >   {
> >   	const struct child_device_config *child;
> >   	u8 dvo_port;
> >   	int i;
> > -	for (i = 0; i < dev_priv->vbt.child_dev_num; i++) {
> > -		child = dev_priv->vbt.child_dev + i;
> > +	for (i = 0; i < i915->vbt.child_dev_num; i++) {
> > +		child = i915->vbt.child_dev + i;
> >   		if (!(child->device_type & DEVICE_TYPE_MIPI_OUTPUT))
> >   			continue;
> > @@ -2152,8 +2152,8 @@ bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv,
> >   		dvo_port = child->dvo_port;
> >   		if (dvo_port == DVO_PORT_MIPIA ||
> > -		    (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(dev_priv) >= 11) ||
> > -		    (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(dev_priv) < 11)) {
> > +		    (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(i915) >= 11) ||
> > +		    (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(i915) < 11)) {
> >   			if (port)
> >   				*port = dvo_port - DVO_PORT_MIPIA;
> >   			return true;
> > @@ -2205,11 +2205,11 @@ intel_bios_is_lspcon_present(const struct drm_i915_private *i915,
> >   	return HAS_LSPCON(i915) && child && child->lspcon;
> >   }
> > -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv,
> > +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915,
> >   				   enum port port)
> >   {
> >   	const struct ddi_vbt_port_info *info =
> > -		&dev_priv->vbt.ddi_port_info[port];
> > +		&i915->vbt.ddi_port_info[port];
> >   	enum aux_ch aux_ch;
> >   	if (!info->alternate_aux_channel) {
> > diff --git a/drivers/gpu/drm/i915/intel_bios.h b/drivers/gpu/drm/i915/intel_bios.h
> > index 4e42cfaf61a7..d292bd1790ae 100644
> > --- a/drivers/gpu/drm/i915/intel_bios.h
> > +++ b/drivers/gpu/drm/i915/intel_bios.h
> > @@ -226,19 +226,19 @@ struct mipi_pps_data {
> >   	u16 panel_power_cycle_delay;
> >   } __packed;
> > -void intel_bios_init(struct drm_i915_private *dev_priv);
> > -void intel_bios_cleanup(struct drm_i915_private *dev_priv);
> > +void intel_bios_init(struct drm_i915_private *i915);
> > +void intel_bios_cleanup(struct drm_i915_private *i915);
> >   bool intel_bios_is_valid_vbt(const void *buf, size_t size);
> > -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv);
> > -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin);
> > -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port);
> > -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port);
> > -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv, enum port port);
> > -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv, enum port *port);
> > +bool intel_bios_is_tv_present(struct drm_i915_private *i915);
> > +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin);
> > +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port);
> > +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port);
> > +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915, enum port port);
> > +bool intel_bios_is_dsi_present(struct drm_i915_private *i915, enum port *port);
> >   bool intel_bios_is_port_hpd_inverted(const struct drm_i915_private *i915,
> >   				     enum port port);
> >   bool intel_bios_is_lspcon_present(const struct drm_i915_private *i915,
> >   				  enum port port);
> > -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv, enum port port);
> > +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915, enum port port);
> >   #endif /* _INTEL_BIOS_H_ */
> > diff --git a/drivers/gpu/drm/i915/intel_bw.c b/drivers/gpu/drm/i915/intel_bw.c
> > index 753ac3165061..1aca44a258d3 100644
> > --- a/drivers/gpu/drm/i915/intel_bw.c
> > +++ b/drivers/gpu/drm/i915/intel_bw.c
> > @@ -22,13 +22,13 @@ struct intel_qgv_info {
> >   	enum intel_dram_type dram_type;
> >   };
> > -static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv,
> > +static int icl_pcode_read_mem_global_info(struct drm_i915_private *i915,
> >   					  struct intel_qgv_info *qi)
> >   {
> >   	u32 val = 0;
> >   	int ret;
> > -	ret = sandybridge_pcode_read(dev_priv,
> > +	ret = sandybridge_pcode_read(i915,
> >   				     ICL_PCODE_MEM_SUBSYSYSTEM_INFO |
> >   				     ICL_PCODE_MEM_SS_READ_GLOBAL_INFO,
> >   				     &val, NULL);
> > @@ -61,14 +61,14 @@ static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv,
> > +static int icl_pcode_read_qgv_point_info(struct drm_i915_private *i915,
> >   					 struct intel_qgv_point *sp,
> >   					 int point)
> >   {
> >   	u32 val = 0, val2;
> >   	int ret;
> > -	ret = sandybridge_pcode_read(dev_priv,
> > +	ret = sandybridge_pcode_read(i915,
> >   				     ICL_PCODE_MEM_SUBSYSYSTEM_INFO |
> >   				     ICL_PCODE_MEM_SS_READ_QGV_POINT_INFO(point),
> >   				     &val, &val2);
> > @@ -87,12 +87,12 @@ static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static int icl_get_qgv_points(struct drm_i915_private *dev_priv,
> > +static int icl_get_qgv_points(struct drm_i915_private *i915,
> >   			      struct intel_qgv_info *qi)
> >   {
> >   	int i, ret;
> > -	ret = icl_pcode_read_mem_global_info(dev_priv, qi);
> > +	ret = icl_pcode_read_mem_global_info(i915, qi);
> >   	if (ret)
> >   		return ret;
> > @@ -102,7 +102,7 @@ static int icl_get_qgv_points(struct drm_i915_private *dev_priv,
> >   	for (i = 0; i < qi->num_points; i++) {
> >   		struct intel_qgv_point *sp = &qi->points[i];
> > -		ret = icl_pcode_read_qgv_point_info(dev_priv, sp, i);
> > +		ret = icl_pcode_read_qgv_point_info(i915, sp, i);
> >   		if (ret)
> >   			return ret;
> > @@ -142,7 +142,7 @@ static const struct intel_sa_info icl_sa_info = {
> >   	.displayrtids = 128,
> >   };
> > -static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> > +static int icl_get_bw_info(struct drm_i915_private *i915)
> >   {
> >   	struct intel_qgv_info qi = {};
> >   	const struct intel_sa_info *sa = &icl_sa_info;
> > @@ -154,7 +154,7 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> >   	int maxdebw;
> >   	int i, ret;
> > -	ret = icl_get_qgv_points(dev_priv, &qi);
> > +	ret = icl_get_qgv_points(i915, &qi);
> >   	if (ret) {
> >   		DRM_DEBUG_KMS("Failed to get memory subsystem information, ignoring bandwidth limits");
> >   		return ret;
> > @@ -170,8 +170,8 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> >   		      icl_calc_bw(dclk_max, 16, 1) * 6 / 10); /* 60% */
> >   	ipqdepth = min(ipqdepthpch, sa->displayrtids / num_channels);
> > -	for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) {
> > -		struct intel_bw_info *bi = &dev_priv->max_bw[i];
> > +	for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) {
> > +		struct intel_bw_info *bi = &i915->max_bw[i];
> >   		int clpchgroup;
> >   		int j;
> > @@ -206,18 +206,18 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv)
> >   	return 0;
> >   }
> > -static unsigned int icl_max_bw(struct drm_i915_private *dev_priv,
> > +static unsigned int icl_max_bw(struct drm_i915_private *i915,
> >   			       int num_planes, int qgv_point)
> >   {
> >   	int i;
> >   	/* Did we initialize the bw limits successfully? */
> > -	if (dev_priv->max_bw[0].num_planes == 0)
> > +	if (i915->max_bw[0].num_planes == 0)
> >   		return UINT_MAX;
> > -	for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) {
> > +	for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) {
> >   		const struct intel_bw_info *bi =
> > -			&dev_priv->max_bw[i];
> > +			&i915->max_bw[i];
> >   		if (num_planes >= bi->num_planes)
> >   			return bi->deratedbw[qgv_point];
> > @@ -226,24 +226,24 @@ static unsigned int icl_max_bw(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -void intel_bw_init_hw(struct drm_i915_private *dev_priv)
> > +void intel_bw_init_hw(struct drm_i915_private *i915)
> >   {
> > -	if (IS_GEN(dev_priv, 11))
> > -		icl_get_bw_info(dev_priv);
> > +	if (IS_GEN(i915, 11))
> > +		icl_get_bw_info(i915);
> >   }
> > -static unsigned int intel_max_data_rate(struct drm_i915_private *dev_priv,
> > +static unsigned int intel_max_data_rate(struct drm_i915_private *i915,
> >   					int num_planes)
> >   {
> > -	if (IS_GEN(dev_priv, 11))
> > +	if (IS_GEN(i915, 11))
> >   		/*
> >   		 * FIXME with SAGV disabled maybe we can assume
> >   		 * point 1 will always be used? Seems to match
> >   		 * the behaviour observed in the wild.
> >   		 */
> > -		return min3(icl_max_bw(dev_priv, num_planes, 0),
> > -			    icl_max_bw(dev_priv, num_planes, 1),
> > -			    icl_max_bw(dev_priv, num_planes, 2));
> > +		return min3(icl_max_bw(i915, num_planes, 0),
> > +			    icl_max_bw(i915, num_planes, 1),
> > +			    icl_max_bw(i915, num_planes, 2));
> >   	else
> >   		return UINT_MAX;
> >   }
> > @@ -293,25 +293,25 @@ void intel_bw_crtc_update(struct intel_bw_state *bw_state,
> >   		      bw_state->num_active_planes[crtc->pipe]);
> >   }
> > -static unsigned int intel_bw_num_active_planes(struct drm_i915_private *dev_priv,
> > +static unsigned int intel_bw_num_active_planes(struct drm_i915_private *i915,
> >   					       const struct intel_bw_state *bw_state)
> >   {
> >   	unsigned int num_active_planes = 0;
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		num_active_planes += bw_state->num_active_planes[pipe];
> >   	return num_active_planes;
> >   }
> > -static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv,
> > +static unsigned int intel_bw_data_rate(struct drm_i915_private *i915,
> >   				       const struct intel_bw_state *bw_state)
> >   {
> >   	unsigned int data_rate = 0;
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		data_rate += bw_state->data_rate[pipe];
> >   	return data_rate;
> > @@ -319,7 +319,7 @@ static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv,
> >   int intel_bw_atomic_check(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc_state *new_crtc_state, *old_crtc_state;
> >   	struct intel_bw_state *bw_state = NULL;
> >   	unsigned int data_rate, max_data_rate;
> > @@ -328,7 +328,7 @@ int intel_bw_atomic_check(struct intel_atomic_state *state)
> >   	int i;
> >   	/* FIXME earlier gens need some checks too */
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return 0;
> >   	for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> > @@ -366,10 +366,10 @@ int intel_bw_atomic_check(struct intel_atomic_state *state)
> >   	if (!bw_state)
> >   		return 0;
> > -	data_rate = intel_bw_data_rate(dev_priv, bw_state);
> > -	num_active_planes = intel_bw_num_active_planes(dev_priv, bw_state);
> > +	data_rate = intel_bw_data_rate(i915, bw_state);
> > +	num_active_planes = intel_bw_num_active_planes(i915, bw_state);
> > -	max_data_rate = intel_max_data_rate(dev_priv, num_active_planes);
> > +	max_data_rate = intel_max_data_rate(i915, num_active_planes);
> >   	data_rate = DIV_ROUND_UP(data_rate, 1000);
> > @@ -406,7 +406,7 @@ static const struct drm_private_state_funcs intel_bw_funcs = {
> >   	.atomic_destroy_state = intel_bw_destroy_state,
> >   };
> > -int intel_bw_init(struct drm_i915_private *dev_priv)
> > +int intel_bw_init(struct drm_i915_private *i915)
> >   {
> >   	struct intel_bw_state *state;
> > @@ -414,7 +414,7 @@ int intel_bw_init(struct drm_i915_private *dev_priv)
> >   	if (!state)
> >   		return -ENOMEM;
> > -	drm_atomic_private_obj_init(&dev_priv->drm, &dev_priv->bw_obj,
> > +	drm_atomic_private_obj_init(&i915->drm, &i915->bw_obj,
> >   				    &state->base, &intel_bw_funcs);
> >   	return 0;
> > diff --git a/drivers/gpu/drm/i915/intel_bw.h b/drivers/gpu/drm/i915/intel_bw.h
> > index e9d9c6d63bc3..396e5029f9a4 100644
> > --- a/drivers/gpu/drm/i915/intel_bw.h
> > +++ b/drivers/gpu/drm/i915/intel_bw.h
> > @@ -27,19 +27,19 @@ struct intel_bw_state {
> >   static inline struct intel_bw_state *
> >   intel_atomic_get_bw_state(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct drm_private_state *bw_state;
> >   	bw_state = drm_atomic_get_private_obj_state(&state->base,
> > -						    &dev_priv->bw_obj);
> > +						    &i915->bw_obj);
> >   	if (IS_ERR(bw_state))
> >   		return ERR_CAST(bw_state);
> >   	return to_intel_bw_state(bw_state);
> >   }
> > -void intel_bw_init_hw(struct drm_i915_private *dev_priv);
> > -int intel_bw_init(struct drm_i915_private *dev_priv);
> > +void intel_bw_init_hw(struct drm_i915_private *i915);
> > +int intel_bw_init(struct drm_i915_private *i915);
> >   int intel_bw_atomic_check(struct intel_atomic_state *state);
> >   void intel_bw_crtc_update(struct intel_bw_state *bw_state,
> >   			  const struct intel_crtc_state *crtc_state);
> > diff --git a/drivers/gpu/drm/i915/intel_cdclk.c b/drivers/gpu/drm/i915/intel_cdclk.c
> > index 8993ab283562..02c617180c4b 100644
> > --- a/drivers/gpu/drm/i915/intel_cdclk.c
> > +++ b/drivers/gpu/drm/i915/intel_cdclk.c
> > @@ -53,46 +53,46 @@
> >    * dividers can be programmed correctly.
> >    */
> > -static void fixed_133mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_133mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 133333;
> >   }
> > -static void fixed_200mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_200mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 200000;
> >   }
> > -static void fixed_266mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_266mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 266667;
> >   }
> > -static void fixed_333mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_333mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 333333;
> >   }
> > -static void fixed_400mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_400mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 400000;
> >   }
> > -static void fixed_450mhz_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void fixed_450mhz_get_cdclk(struct drm_i915_private *i915,
> >   				   struct intel_cdclk_state *cdclk_state)
> >   {
> >   	cdclk_state->cdclk = 450000;
> >   }
> > -static void i85x_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void i85x_get_cdclk(struct drm_i915_private *i915,
> >   			   struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u16 hpllcc = 0;
> >   	/*
> > @@ -131,10 +131,10 @@ static void i85x_get_cdclk(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void i915gm_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void i915gm_get_cdclk(struct drm_i915_private *i915,
> >   			     struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u16 gcfgc = 0;
> >   	pci_read_config_word(pdev, GCFGC, &gcfgc);
> > @@ -155,10 +155,10 @@ static void i915gm_get_cdclk(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void i945gm_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void i945gm_get_cdclk(struct drm_i915_private *i915,
> >   			     struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u16 gcfgc = 0;
> >   	pci_read_config_word(pdev, GCFGC, &gcfgc);
> > @@ -179,7 +179,7 @@ static void i945gm_get_cdclk(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> > +static unsigned int intel_hpll_vco(struct drm_i915_private *i915)
> >   {
> >   	static const unsigned int blb_vco[8] = {
> >   		[0] = 3200000,
> > @@ -223,20 +223,20 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> >   	u8 tmp = 0;
> >   	/* FIXME other chipsets? */
> > -	if (IS_GM45(dev_priv))
> > +	if (IS_GM45(i915))
> >   		vco_table = ctg_vco;
> > -	else if (IS_G45(dev_priv))
> > +	else if (IS_G45(i915))
> >   		vco_table = elk_vco;
> > -	else if (IS_I965GM(dev_priv))
> > +	else if (IS_I965GM(i915))
> >   		vco_table = cl_vco;
> > -	else if (IS_PINEVIEW(dev_priv))
> > +	else if (IS_PINEVIEW(i915))
> >   		vco_table = pnv_vco;
> > -	else if (IS_G33(dev_priv))
> > +	else if (IS_G33(i915))
> >   		vco_table = blb_vco;
> >   	else
> >   		return 0;
> > -	tmp = I915_READ(IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv) ?
> > +	tmp = I915_READ(IS_PINEVIEW(i915) || IS_MOBILE(i915) ?
> >   			HPLLVCO_MOBILE : HPLLVCO);
> >   	vco = vco_table[tmp & 0x7];
> > @@ -248,10 +248,10 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv)
> >   	return vco;
> >   }
> > -static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void g33_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	static const u8 div_3200[] = { 12, 10,  8,  7, 5, 16 };
> >   	static const u8 div_4000[] = { 14, 12, 10,  8, 6, 20 };
> >   	static const u8 div_4800[] = { 20, 14, 12, 10, 8, 24 };
> > @@ -260,7 +260,7 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> >   	unsigned int cdclk_sel;
> >   	u16 tmp = 0;
> > -	cdclk_state->vco = intel_hpll_vco(dev_priv);
> > +	cdclk_state->vco = intel_hpll_vco(i915);
> >   	pci_read_config_word(pdev, GCFGC, &tmp);
> > @@ -296,10 +296,10 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv,
> >   	cdclk_state->cdclk = 190476;
> >   }
> > -static void pnv_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void pnv_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u16 gcfgc = 0;
> >   	pci_read_config_word(pdev, GCFGC, &gcfgc);
> > @@ -329,10 +329,10 @@ static void pnv_get_cdclk(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void i965gm_get_cdclk(struct drm_i915_private *i915,
> >   			     struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	static const u8 div_3200[] = { 16, 10,  8 };
> >   	static const u8 div_4000[] = { 20, 12, 10 };
> >   	static const u8 div_5333[] = { 24, 16, 14 };
> > @@ -340,7 +340,7 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> >   	unsigned int cdclk_sel;
> >   	u16 tmp = 0;
> > -	cdclk_state->vco = intel_hpll_vco(dev_priv);
> > +	cdclk_state->vco = intel_hpll_vco(i915);
> >   	pci_read_config_word(pdev, GCFGC, &tmp);
> > @@ -373,14 +373,14 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv,
> >   	cdclk_state->cdclk = 200000;
> >   }
> > -static void gm45_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void gm45_get_cdclk(struct drm_i915_private *i915,
> >   			   struct intel_cdclk_state *cdclk_state)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	unsigned int cdclk_sel;
> >   	u16 tmp = 0;
> > -	cdclk_state->vco = intel_hpll_vco(dev_priv);
> > +	cdclk_state->vco = intel_hpll_vco(i915);
> >   	pci_read_config_word(pdev, GCFGC, &tmp);
> > @@ -403,7 +403,7 @@ static void gm45_get_cdclk(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void hsw_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void hsw_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 lcpll = I915_READ(LCPLL_CTL);
> > @@ -415,15 +415,15 @@ static void hsw_get_cdclk(struct drm_i915_private *dev_priv,
> >   		cdclk_state->cdclk = 450000;
> >   	else if (freq == LCPLL_CLK_FREQ_450)
> >   		cdclk_state->cdclk = 450000;
> > -	else if (IS_HSW_ULT(dev_priv))
> > +	else if (IS_HSW_ULT(i915))
> >   		cdclk_state->cdclk = 337500;
> >   	else
> >   		cdclk_state->cdclk = 540000;
> >   }
> > -static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> > +static int vlv_calc_cdclk(struct drm_i915_private *i915, int min_cdclk)
> >   {
> > -	int freq_320 = (dev_priv->hpll_freq <<  1) % 320000 != 0 ?
> > +	int freq_320 = (i915->hpll_freq <<  1) % 320000 != 0 ?
> >   		333333 : 320000;
> >   	/*
> > @@ -431,7 +431,7 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> >   	 * Not sure what's wrong. For now use 200MHz only when all pipes
> >   	 * are off.
> >   	 */
> > -	if (IS_VALLEYVIEW(dev_priv) && min_cdclk > freq_320)
> > +	if (IS_VALLEYVIEW(i915) && min_cdclk > freq_320)
> >   		return 400000;
> >   	else if (min_cdclk > 266667)
> >   		return freq_320;
> > @@ -441,9 +441,9 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk)
> >   		return 200000;
> >   }
> > -static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk)
> > +static u8 vlv_calc_voltage_level(struct drm_i915_private *i915, int cdclk)
> >   {
> > -	if (IS_VALLEYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915)) {
> >   		if (cdclk >= 320000) /* jump to highest voltage for 400MHz too */
> >   			return 2;
> >   		else if (cdclk >= 266667)
> > @@ -456,29 +456,29 @@ static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk)
> >   		 * hardware has shown that we just need to write the desired
> >   		 * CCK divider into the Punit register.
> >   		 */
> > -		return DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1, cdclk) - 1;
> > +		return DIV_ROUND_CLOSEST(i915->hpll_freq << 1, cdclk) - 1;
> >   	}
> >   }
> > -static void vlv_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void vlv_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 val;
> > -	vlv_iosf_sb_get(dev_priv,
> > +	vlv_iosf_sb_get(i915,
> >   			BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT));
> > -	cdclk_state->vco = vlv_get_hpll_vco(dev_priv);
> > -	cdclk_state->cdclk = vlv_get_cck_clock(dev_priv, "cdclk",
> > +	cdclk_state->vco = vlv_get_hpll_vco(i915);
> > +	cdclk_state->cdclk = vlv_get_cck_clock(i915, "cdclk",
> >   					       CCK_DISPLAY_CLOCK_CONTROL,
> >   					       cdclk_state->vco);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +	val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> > -	vlv_iosf_sb_put(dev_priv,
> > +	vlv_iosf_sb_put(i915,
> >   			BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT));
> > -	if (IS_VALLEYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915))
> >   		cdclk_state->voltage_level = (val & DSPFREQGUAR_MASK) >>
> >   			DSPFREQGUAR_SHIFT;
> >   	else
> > @@ -486,18 +486,18 @@ static void vlv_get_cdclk(struct drm_i915_private *dev_priv,
> >   			DSPFREQGUAR_SHIFT_CHV;
> >   }
> > -static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv)
> > +static void vlv_program_pfi_credits(struct drm_i915_private *i915)
> >   {
> >   	unsigned int credits, default_credits;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		default_credits = PFI_CREDIT(12);
> >   	else
> >   		default_credits = PFI_CREDIT(8);
> > -	if (dev_priv->cdclk.hw.cdclk >= dev_priv->czclk_freq) {
> > +	if (i915->cdclk.hw.cdclk >= i915->czclk_freq) {
> >   		/* CHV suggested value is 31 or 63 */
> > -		if (IS_CHERRYVIEW(dev_priv))
> > +		if (IS_CHERRYVIEW(i915))
> >   			credits = PFI_CREDIT_63;
> >   		else
> >   			credits = PFI_CREDIT(15);
> > @@ -522,7 +522,7 @@ static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv)
> >   	WARN_ON(I915_READ(GCI_CONTROL) & PFI_CREDIT_RESEND);
> >   }
> > -static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void vlv_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -548,18 +548,18 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> >   	 * a system suspend.  So grab the PIPE-A domain, which covers
> >   	 * the HW blocks needed for the following programming.
> >   	 */
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A);
> > -	vlv_iosf_sb_get(dev_priv,
> > +	vlv_iosf_sb_get(i915,
> >   			BIT(VLV_IOSF_SB_CCK) |
> >   			BIT(VLV_IOSF_SB_BUNIT) |
> >   			BIT(VLV_IOSF_SB_PUNIT));
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +	val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> >   	val &= ~DSPFREQGUAR_MASK;
> >   	val |= (cmd << DSPFREQGUAR_SHIFT);
> > -	vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> > -	if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) &
> > +	vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
> > +	if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) &
> >   		      DSPFREQSTAT_MASK) == (cmd << DSPFREQSTAT_SHIFT),
> >   		     50)) {
> >   		DRM_ERROR("timed out waiting for CDclk change\n");
> > @@ -568,23 +568,23 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> >   	if (cdclk == 400000) {
> >   		u32 divider;
> > -		divider = DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1,
> > +		divider = DIV_ROUND_CLOSEST(i915->hpll_freq << 1,
> >   					    cdclk) - 1;
> >   		/* adjust cdclk divider */
> > -		val = vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL);
> > +		val = vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL);
> >   		val &= ~CCK_FREQUENCY_VALUES;
> >   		val |= divider;
> > -		vlv_cck_write(dev_priv, CCK_DISPLAY_CLOCK_CONTROL, val);
> > +		vlv_cck_write(i915, CCK_DISPLAY_CLOCK_CONTROL, val);
> > -		if (wait_for((vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL) &
> > +		if (wait_for((vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL) &
> >   			      CCK_FREQUENCY_STATUS) == (divider << CCK_FREQUENCY_STATUS_SHIFT),
> >   			     50))
> >   			DRM_ERROR("timed out waiting for CDclk change\n");
> >   	}
> >   	/* adjust self-refresh exit latency value */
> > -	val = vlv_bunit_read(dev_priv, BUNIT_REG_BISOC);
> > +	val = vlv_bunit_read(i915, BUNIT_REG_BISOC);
> >   	val &= ~0x7f;
> >   	/*
> > @@ -595,21 +595,21 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv,
> >   		val |= 4500 / 250; /* 4.5 usec */
> >   	else
> >   		val |= 3000 / 250; /* 3.0 usec */
> > -	vlv_bunit_write(dev_priv, BUNIT_REG_BISOC, val);
> > +	vlv_bunit_write(i915, BUNIT_REG_BISOC, val);
> > -	vlv_iosf_sb_put(dev_priv,
> > +	vlv_iosf_sb_put(i915,
> >   			BIT(VLV_IOSF_SB_CCK) |
> >   			BIT(VLV_IOSF_SB_BUNIT) |
> >   			BIT(VLV_IOSF_SB_PUNIT));
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> > -	vlv_program_pfi_credits(dev_priv);
> > +	vlv_program_pfi_credits(i915);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref);
> >   }
> > -static void chv_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void chv_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -634,26 +634,26 @@ static void chv_set_cdclk(struct drm_i915_private *dev_priv,
> >   	 * a system suspend.  So grab the PIPE-A domain, which covers
> >   	 * the HW blocks needed for the following programming.
> >   	 */
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A);
> > -	vlv_punit_get(dev_priv);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +	vlv_punit_get(i915);
> > +	val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> >   	val &= ~DSPFREQGUAR_MASK_CHV;
> >   	val |= (cmd << DSPFREQGUAR_SHIFT_CHV);
> > -	vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> > -	if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) &
> > +	vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
> > +	if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) &
> >   		      DSPFREQSTAT_MASK_CHV) == (cmd << DSPFREQSTAT_SHIFT_CHV),
> >   		     50)) {
> >   		DRM_ERROR("timed out waiting for CDclk change\n");
> >   	}
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> > -	vlv_program_pfi_credits(dev_priv);
> > +	vlv_program_pfi_credits(i915);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref);
> >   }
> >   static int bdw_calc_cdclk(int min_cdclk)
> > @@ -683,7 +683,7 @@ static u8 bdw_calc_voltage_level(int cdclk)
> >   	}
> >   }
> > -static void bdw_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void bdw_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 lcpll = I915_READ(LCPLL_CTL);
> > @@ -710,7 +710,7 @@ static void bdw_get_cdclk(struct drm_i915_private *dev_priv,
> >   		bdw_calc_voltage_level(cdclk_state->cdclk);
> >   }
> > -static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void bdw_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -726,7 +726,7 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> >   		 "trying to change cdclk frequency with cdclk not enabled\n"))
> >   		return;
> > -	ret = sandybridge_pcode_write(dev_priv,
> > +	ret = sandybridge_pcode_write(i915,
> >   				      BDW_PCODE_DISPLAY_FREQ_CHANGE_REQ, 0x0);
> >   	if (ret) {
> >   		DRM_ERROR("failed to inform pcode about cdclk change\n");
> > @@ -776,12 +776,12 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv,
> >   			LCPLL_CD_SOURCE_FCLK_DONE) == 0, 1))
> >   		DRM_ERROR("Switching back to LCPLL failed\n");
> > -	sandybridge_pcode_write(dev_priv, HSW_PCODE_DE_WRITE_FREQ_REQ,
> > +	sandybridge_pcode_write(i915, HSW_PCODE_DE_WRITE_FREQ_REQ,
> >   				cdclk_state->voltage_level);
> >   	I915_WRITE(CDCLK_FREQ, DIV_ROUND_CLOSEST(cdclk, 1000) - 1);
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> >   }
> >   static int skl_calc_cdclk(int min_cdclk, int vco)
> > @@ -819,7 +819,7 @@ static u8 skl_calc_voltage_level(int cdclk)
> >   		return 0;
> >   }
> > -static void skl_dpll0_update(struct drm_i915_private *dev_priv,
> > +static void skl_dpll0_update(struct drm_i915_private *i915,
> >   			     struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 val;
> > @@ -859,12 +859,12 @@ static void skl_dpll0_update(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void skl_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void skl_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 cdctl;
> > -	skl_dpll0_update(dev_priv, cdclk_state);
> > +	skl_dpll0_update(i915, cdclk_state);
> >   	cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
> > @@ -926,18 +926,18 @@ static int skl_cdclk_decimal(int cdclk)
> >   	return DIV_ROUND_CLOSEST(cdclk - 1000, 500);
> >   }
> > -static void skl_set_preferred_cdclk_vco(struct drm_i915_private *dev_priv,
> > +static void skl_set_preferred_cdclk_vco(struct drm_i915_private *i915,
> >   					int vco)
> >   {
> > -	bool changed = dev_priv->skl_preferred_vco_freq != vco;
> > +	bool changed = i915->skl_preferred_vco_freq != vco;
> > -	dev_priv->skl_preferred_vco_freq = vco;
> > +	i915->skl_preferred_vco_freq = vco;
> >   	if (changed)
> > -		intel_update_max_cdclk(dev_priv);
> > +		intel_update_max_cdclk(i915);
> >   }
> > -static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco)
> > +static void skl_dpll0_enable(struct drm_i915_private *i915, int vco)
> >   {
> >   	u32 val;
> > @@ -969,29 +969,29 @@ static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco)
> >   	I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) | LCPLL_PLL_ENABLE);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    LCPLL1_CTL, LCPLL_PLL_LOCK, LCPLL_PLL_LOCK,
> >   				    5))
> >   		DRM_ERROR("DPLL0 not locked\n");
> > -	dev_priv->cdclk.hw.vco = vco;
> > +	i915->cdclk.hw.vco = vco;
> >   	/* We'll want to keep using the current vco from now on. */
> > -	skl_set_preferred_cdclk_vco(dev_priv, vco);
> > +	skl_set_preferred_cdclk_vco(i915, vco);
> >   }
> > -static void skl_dpll0_disable(struct drm_i915_private *dev_priv)
> > +static void skl_dpll0_disable(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) & ~LCPLL_PLL_ENABLE);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    LCPLL1_CTL, LCPLL_PLL_LOCK, 0,
> >   				    1))
> >   		DRM_ERROR("Couldn't disable DPLL0\n");
> > -	dev_priv->cdclk.hw.vco = 0;
> > +	i915->cdclk.hw.vco = 0;
> >   }
> > -static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void skl_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -1008,9 +1008,9 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	 * use the corresponding VCO freq as that always leads to using the
> >   	 * minimum 308MHz CDCLK.
> >   	 */
> > -	WARN_ON_ONCE(IS_SKYLAKE(dev_priv) && vco == 8640000);
> > +	WARN_ON_ONCE(IS_SKYLAKE(i915) && vco == 8640000);
> > -	ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				SKL_CDCLK_PREPARE_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE, 3);
> > @@ -1023,7 +1023,7 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	/* Choose frequency for this cdclk */
> >   	switch (cdclk) {
> >   	default:
> > -		WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> > +		WARN_ON(cdclk != i915->cdclk.hw.bypass);
> >   		WARN_ON(vco != 0);
> >   		/* fall through */
> >   	case 308571:
> > @@ -1043,13 +1043,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> >   		break;
> >   	}
> > -	if (dev_priv->cdclk.hw.vco != 0 &&
> > -	    dev_priv->cdclk.hw.vco != vco)
> > -		skl_dpll0_disable(dev_priv);
> > +	if (i915->cdclk.hw.vco != 0 &&
> > +	    i915->cdclk.hw.vco != vco)
> > +		skl_dpll0_disable(i915);
> >   	cdclk_ctl = I915_READ(CDCLK_CTL);
> > -	if (dev_priv->cdclk.hw.vco != vco) {
> > +	if (i915->cdclk.hw.vco != vco) {
> >   		/* Wa Display #1183: skl,kbl,cfl */
> >   		cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK);
> >   		cdclk_ctl |= freq_select | skl_cdclk_decimal(cdclk);
> > @@ -1061,8 +1061,8 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CDCLK_CTL, cdclk_ctl);
> >   	POSTING_READ(CDCLK_CTL);
> > -	if (dev_priv->cdclk.hw.vco != vco)
> > -		skl_dpll0_enable(dev_priv, vco);
> > +	if (i915->cdclk.hw.vco != vco)
> > +		skl_dpll0_enable(i915, vco);
> >   	/* Wa Display #1183: skl,kbl,cfl */
> >   	cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK);
> > @@ -1077,13 +1077,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(CDCLK_CTL);
> >   	/* inform PCU of the change */
> > -	sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				cdclk_state->voltage_level);
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> >   }
> > -static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> > +static void skl_sanitize_cdclk(struct drm_i915_private *i915)
> >   {
> >   	u32 cdctl, expected;
> > @@ -1095,12 +1095,12 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	if ((I915_READ(SWF_ILK(0x18)) & 0x00FFFFFF) == 0)
> >   		goto sanitize;
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> >   	/* Is PLL enabled and locked ? */
> > -	if (dev_priv->cdclk.hw.vco == 0 ||
> > -	    dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (i915->cdclk.hw.vco == 0 ||
> > +	    i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> >   		goto sanitize;
> >   	/* DPLL okay; verify the cdclock
> > @@ -1111,7 +1111,7 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	 */
> >   	cdctl = I915_READ(CDCLK_CTL);
> >   	expected = (cdctl & CDCLK_FREQ_SEL_MASK) |
> > -		skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> > +		skl_cdclk_decimal(i915->cdclk.hw.cdclk);
> >   	if (cdctl == expected)
> >   		/* All well; nothing to sanitize */
> >   		return;
> > @@ -1120,49 +1120,49 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
> >   	/* force cdclk programming */
> > -	dev_priv->cdclk.hw.cdclk = 0;
> > +	i915->cdclk.hw.cdclk = 0;
> >   	/* force full PLL disable + enable */
> > -	dev_priv->cdclk.hw.vco = -1;
> > +	i915->cdclk.hw.vco = -1;
> >   }
> > -static void skl_init_cdclk(struct drm_i915_private *dev_priv)
> > +static void skl_init_cdclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_cdclk_state cdclk_state;
> > -	skl_sanitize_cdclk(dev_priv);
> > +	skl_sanitize_cdclk(i915);
> > -	if (dev_priv->cdclk.hw.cdclk != 0 &&
> > -	    dev_priv->cdclk.hw.vco != 0) {
> > +	if (i915->cdclk.hw.cdclk != 0 &&
> > +	    i915->cdclk.hw.vco != 0) {
> >   		/*
> >   		 * Use the current vco as our initial
> >   		 * guess as to what the preferred vco is.
> >   		 */
> > -		if (dev_priv->skl_preferred_vco_freq == 0)
> > -			skl_set_preferred_cdclk_vco(dev_priv,
> > -						    dev_priv->cdclk.hw.vco);
> > +		if (i915->skl_preferred_vco_freq == 0)
> > +			skl_set_preferred_cdclk_vco(i915,
> > +						    i915->cdclk.hw.vco);
> >   		return;
> >   	}
> > -	cdclk_state = dev_priv->cdclk.hw;
> > +	cdclk_state = i915->cdclk.hw;
> > -	cdclk_state.vco = dev_priv->skl_preferred_vco_freq;
> > +	cdclk_state.vco = i915->skl_preferred_vco_freq;
> >   	if (cdclk_state.vco == 0)
> >   		cdclk_state.vco = 8100000;
> >   	cdclk_state.cdclk = skl_calc_cdclk(0, cdclk_state.vco);
> >   	cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk);
> > -	skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> > -static void skl_uninit_cdclk(struct drm_i915_private *dev_priv)
> > +static void skl_uninit_cdclk(struct drm_i915_private *i915)
> >   {
> > -	struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> > +	struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
> >   	cdclk_state.cdclk = cdclk_state.bypass;
> >   	cdclk_state.vco = 0;
> >   	cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk);
> > -	skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> >   static int bxt_calc_cdclk(int min_cdclk)
> > @@ -1194,11 +1194,11 @@ static u8 bxt_calc_voltage_level(int cdclk)
> >   	return DIV_ROUND_UP(cdclk, 25000);
> >   }
> > -static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> > +static int bxt_de_pll_vco(struct drm_i915_private *i915, int cdclk)
> >   {
> >   	int ratio;
> > -	if (cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (cdclk == i915->cdclk.hw.bypass)
> >   		return 0;
> >   	switch (cdclk) {
> > @@ -1216,14 +1216,14 @@ static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> >   		break;
> >   	}
> > -	return dev_priv->cdclk.hw.ref * ratio;
> > +	return i915->cdclk.hw.ref * ratio;
> >   }
> > -static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> > +static int glk_de_pll_vco(struct drm_i915_private *i915, int cdclk)
> >   {
> >   	int ratio;
> > -	if (cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (cdclk == i915->cdclk.hw.bypass)
> >   		return 0;
> >   	switch (cdclk) {
> > @@ -1237,10 +1237,10 @@ static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> >   		break;
> >   	}
> > -	return dev_priv->cdclk.hw.ref * ratio;
> > +	return i915->cdclk.hw.ref * ratio;
> >   }
> > -static void bxt_de_pll_update(struct drm_i915_private *dev_priv,
> > +static void bxt_de_pll_update(struct drm_i915_private *i915,
> >   			      struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 val;
> > @@ -1259,13 +1259,13 @@ static void bxt_de_pll_update(struct drm_i915_private *dev_priv,
> >   	cdclk_state->vco = (val & BXT_DE_PLL_RATIO_MASK) * cdclk_state->ref;
> >   }
> > -static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void bxt_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 divider;
> >   	int div;
> > -	bxt_de_pll_update(dev_priv, cdclk_state);
> > +	bxt_de_pll_update(i915, cdclk_state);
> >   	cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
> > @@ -1279,7 +1279,7 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> >   		div = 2;
> >   		break;
> >   	case BXT_CDCLK_CD2X_DIV_SEL_1_5:
> > -		WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n");
> > +		WARN(IS_GEMINILAKE(i915), "Unsupported divider\n");
> >   		div = 3;
> >   		break;
> >   	case BXT_CDCLK_CD2X_DIV_SEL_2:
> > @@ -1304,22 +1304,22 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv,
> >   		bxt_calc_voltage_level(cdclk_state->cdclk);
> >   }
> > -static void bxt_de_pll_disable(struct drm_i915_private *dev_priv)
> > +static void bxt_de_pll_disable(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(BXT_DE_PLL_ENABLE, 0);
> >   	/* Timeout 200us */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    BXT_DE_PLL_ENABLE, BXT_DE_PLL_LOCK, 0,
> >   				    1))
> >   		DRM_ERROR("timeout waiting for DE PLL unlock\n");
> > -	dev_priv->cdclk.hw.vco = 0;
> > +	i915->cdclk.hw.vco = 0;
> >   }
> > -static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco)
> > +static void bxt_de_pll_enable(struct drm_i915_private *i915, int vco)
> >   {
> > -	int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref);
> > +	int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref);
> >   	u32 val;
> >   	val = I915_READ(BXT_DE_PLL_CTL);
> > @@ -1330,17 +1330,17 @@ static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco)
> >   	I915_WRITE(BXT_DE_PLL_ENABLE, BXT_DE_PLL_PLL_ENABLE);
> >   	/* Timeout 200us */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    BXT_DE_PLL_ENABLE,
> >   				    BXT_DE_PLL_LOCK,
> >   				    BXT_DE_PLL_LOCK,
> >   				    1))
> >   		DRM_ERROR("timeout waiting for DE PLL lock\n");
> > -	dev_priv->cdclk.hw.vco = vco;
> > +	i915->cdclk.hw.vco = vco;
> >   }
> > -static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void bxt_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -1352,14 +1352,14 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   	/* cdclk = vco / 2 / div{1,1.5,2,4} */
> >   	switch (DIV_ROUND_CLOSEST(vco, cdclk)) {
> >   	default:
> > -		WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> > +		WARN_ON(cdclk != i915->cdclk.hw.bypass);
> >   		WARN_ON(vco != 0);
> >   		/* fall through */
> >   	case 2:
> >   		divider = BXT_CDCLK_CD2X_DIV_SEL_1;
> >   		break;
> >   	case 3:
> > -		WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n");
> > +		WARN(IS_GEMINILAKE(i915), "Unsupported divider\n");
> >   		divider = BXT_CDCLK_CD2X_DIV_SEL_1_5;
> >   		break;
> >   	case 4:
> > @@ -1375,7 +1375,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   	 * requires us to wait up to 150usec, but that leads to timeouts;
> >   	 * the 2ms used here is based on experiment.
> >   	 */
> > -	ret = sandybridge_pcode_write_timeout(dev_priv,
> > +	ret = sandybridge_pcode_write_timeout(i915,
> >   					      HSW_PCODE_DE_WRITE_FREQ_REQ,
> >   					      0x80000000, 150, 2);
> >   	if (ret) {
> > @@ -1384,12 +1384,12 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	if (dev_priv->cdclk.hw.vco != 0 &&
> > -	    dev_priv->cdclk.hw.vco != vco)
> > -		bxt_de_pll_disable(dev_priv);
> > +	if (i915->cdclk.hw.vco != 0 &&
> > +	    i915->cdclk.hw.vco != vco)
> > +		bxt_de_pll_disable(i915);
> > -	if (dev_priv->cdclk.hw.vco != vco)
> > -		bxt_de_pll_enable(dev_priv, vco);
> > +	if (i915->cdclk.hw.vco != vco)
> > +		bxt_de_pll_enable(i915, vco);
> >   	val = divider | skl_cdclk_decimal(cdclk);
> >   	if (pipe == INVALID_PIPE)
> > @@ -1405,7 +1405,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CDCLK_CTL, val);
> >   	if (pipe != INVALID_PIPE)
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> >   	/*
> >   	 * The timeout isn't specified, the 2ms used here is based on
> > @@ -1413,7 +1413,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   	 * FIXME: Waiting for the request completion could be delayed until
> >   	 * the next PCODE request based on BSpec.
> >   	 */
> > -	ret = sandybridge_pcode_write_timeout(dev_priv,
> > +	ret = sandybridge_pcode_write_timeout(i915,
> >   					      HSW_PCODE_DE_WRITE_FREQ_REQ,
> >   					      cdclk_state->voltage_level, 150, 2);
> >   	if (ret) {
> > @@ -1422,18 +1422,18 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> >   }
> > -static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> > +static void bxt_sanitize_cdclk(struct drm_i915_private *i915)
> >   {
> >   	u32 cdctl, expected;
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> > -	if (dev_priv->cdclk.hw.vco == 0 ||
> > -	    dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (i915->cdclk.hw.vco == 0 ||
> > +	    i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> >   		goto sanitize;
> >   	/* DPLL okay; verify the cdclock
> > @@ -1451,12 +1451,12 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE;
> >   	expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) |
> > -		skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> > +		skl_cdclk_decimal(i915->cdclk.hw.cdclk);
> >   	/*
> >   	 * Disable SSA Precharge when CD clock frequency < 500 MHz,
> >   	 * enable otherwise.
> >   	 */
> > -	if (dev_priv->cdclk.hw.cdclk >= 500000)
> > +	if (i915->cdclk.hw.cdclk >= 500000)
> >   		expected |= BXT_CDCLK_SSA_PRECHARGE_ENABLE;
> >   	if (cdctl == expected)
> > @@ -1467,50 +1467,50 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
> >   	/* force cdclk programming */
> > -	dev_priv->cdclk.hw.cdclk = 0;
> > +	i915->cdclk.hw.cdclk = 0;
> >   	/* force full PLL disable + enable */
> > -	dev_priv->cdclk.hw.vco = -1;
> > +	i915->cdclk.hw.vco = -1;
> >   }
> > -static void bxt_init_cdclk(struct drm_i915_private *dev_priv)
> > +static void bxt_init_cdclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_cdclk_state cdclk_state;
> > -	bxt_sanitize_cdclk(dev_priv);
> > +	bxt_sanitize_cdclk(i915);
> > -	if (dev_priv->cdclk.hw.cdclk != 0 &&
> > -	    dev_priv->cdclk.hw.vco != 0)
> > +	if (i915->cdclk.hw.cdclk != 0 &&
> > +	    i915->cdclk.hw.vco != 0)
> >   		return;
> > -	cdclk_state = dev_priv->cdclk.hw;
> > +	cdclk_state = i915->cdclk.hw;
> >   	/*
> >   	 * FIXME:
> >   	 * - The initial CDCLK needs to be read from VBT.
> >   	 *   Need to make this change after VBT has changes for BXT.
> >   	 */
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		cdclk_state.cdclk = glk_calc_cdclk(0);
> > -		cdclk_state.vco = glk_de_pll_vco(dev_priv, cdclk_state.cdclk);
> > +		cdclk_state.vco = glk_de_pll_vco(i915, cdclk_state.cdclk);
> >   	} else {
> >   		cdclk_state.cdclk = bxt_calc_cdclk(0);
> > -		cdclk_state.vco = bxt_de_pll_vco(dev_priv, cdclk_state.cdclk);
> > +		cdclk_state.vco = bxt_de_pll_vco(i915, cdclk_state.cdclk);
> >   	}
> >   	cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk);
> > -	bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> > -static void bxt_uninit_cdclk(struct drm_i915_private *dev_priv)
> > +static void bxt_uninit_cdclk(struct drm_i915_private *i915)
> >   {
> > -	struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> > +	struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
> >   	cdclk_state.cdclk = cdclk_state.bypass;
> >   	cdclk_state.vco = 0;
> >   	cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk);
> > -	bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> >   static int cnl_calc_cdclk(int min_cdclk)
> > @@ -1533,7 +1533,7 @@ static u8 cnl_calc_voltage_level(int cdclk)
> >   		return 0;
> >   }
> > -static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv,
> > +static void cnl_cdclk_pll_update(struct drm_i915_private *i915,
> >   				 struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 val;
> > @@ -1555,13 +1555,13 @@ static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv,
> >   	cdclk_state->vco = (val & CNL_CDCLK_PLL_RATIO_MASK) * cdclk_state->ref;
> >   }
> > -static void cnl_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void cnl_get_cdclk(struct drm_i915_private *i915,
> >   			 struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 divider;
> >   	int div;
> > -	cnl_cdclk_pll_update(dev_priv, cdclk_state);
> > +	cnl_cdclk_pll_update(i915, cdclk_state);
> >   	cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref;
> > @@ -1593,7 +1593,7 @@ static void cnl_get_cdclk(struct drm_i915_private *dev_priv,
> >   		cnl_calc_voltage_level(cdclk_state->cdclk);
> >   }
> > -static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv)
> > +static void cnl_cdclk_pll_disable(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -1605,12 +1605,12 @@ static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv)
> >   	if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) == 0, 1))
> >   		DRM_ERROR("timeout waiting for CDCLK PLL unlock\n");
> > -	dev_priv->cdclk.hw.vco = 0;
> > +	i915->cdclk.hw.vco = 0;
> >   }
> > -static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco)
> > +static void cnl_cdclk_pll_enable(struct drm_i915_private *i915, int vco)
> >   {
> > -	int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref);
> > +	int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref);
> >   	u32 val;
> >   	val = CNL_CDCLK_PLL_RATIO(ratio);
> > @@ -1623,10 +1623,10 @@ static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco)
> >   	if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) != 0, 1))
> >   		DRM_ERROR("timeout waiting for CDCLK PLL lock\n");
> > -	dev_priv->cdclk.hw.vco = vco;
> > +	i915->cdclk.hw.vco = vco;
> >   }
> > -static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void cnl_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -1635,7 +1635,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	u32 val, divider;
> >   	int ret;
> > -	ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				SKL_CDCLK_PREPARE_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE, 3);
> > @@ -1648,7 +1648,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	/* cdclk = vco / 2 / div{1,2} */
> >   	switch (DIV_ROUND_CLOSEST(vco, cdclk)) {
> >   	default:
> > -		WARN_ON(cdclk != dev_priv->cdclk.hw.bypass);
> > +		WARN_ON(cdclk != i915->cdclk.hw.bypass);
> >   		WARN_ON(vco != 0);
> >   		/* fall through */
> >   	case 2:
> > @@ -1659,12 +1659,12 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> >   		break;
> >   	}
> > -	if (dev_priv->cdclk.hw.vco != 0 &&
> > -	    dev_priv->cdclk.hw.vco != vco)
> > -		cnl_cdclk_pll_disable(dev_priv);
> > +	if (i915->cdclk.hw.vco != 0 &&
> > +	    i915->cdclk.hw.vco != vco)
> > +		cnl_cdclk_pll_disable(i915);
> > -	if (dev_priv->cdclk.hw.vco != vco)
> > -		cnl_cdclk_pll_enable(dev_priv, vco);
> > +	if (i915->cdclk.hw.vco != vco)
> > +		cnl_cdclk_pll_enable(i915, vco);
> >   	val = divider | skl_cdclk_decimal(cdclk);
> >   	if (pipe == INVALID_PIPE)
> > @@ -1674,26 +1674,26 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CDCLK_CTL, val);
> >   	if (pipe != INVALID_PIPE)
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> >   	/* inform PCU of the change */
> > -	sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				cdclk_state->voltage_level);
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> >   	/*
> >   	 * Can't read out the voltage level :(
> >   	 * Let's just assume everything is as expected.
> >   	 */
> > -	dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> > +	i915->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> >   }
> > -static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> > +static int cnl_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk)
> >   {
> >   	int ratio;
> > -	if (cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (cdclk == i915->cdclk.hw.bypass)
> >   		return 0;
> >   	switch (cdclk) {
> > @@ -1702,25 +1702,25 @@ static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> >   		/* fall through */
> >   	case 168000:
> >   	case 336000:
> > -		ratio = dev_priv->cdclk.hw.ref == 19200 ? 35 : 28;
> > +		ratio = i915->cdclk.hw.ref == 19200 ? 35 : 28;
> >   		break;
> >   	case 528000:
> > -		ratio = dev_priv->cdclk.hw.ref == 19200 ? 55 : 44;
> > +		ratio = i915->cdclk.hw.ref == 19200 ? 55 : 44;
> >   		break;
> >   	}
> > -	return dev_priv->cdclk.hw.ref * ratio;
> > +	return i915->cdclk.hw.ref * ratio;
> >   }
> > -static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> > +static void cnl_sanitize_cdclk(struct drm_i915_private *i915)
> >   {
> >   	u32 cdctl, expected;
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> > -	if (dev_priv->cdclk.hw.vco == 0 ||
> > -	    dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (i915->cdclk.hw.vco == 0 ||
> > +	    i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> >   		goto sanitize;
> >   	/* DPLL okay; verify the cdclock
> > @@ -1738,7 +1738,7 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE;
> >   	expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) |
> > -		   skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk);
> > +		   skl_cdclk_decimal(i915->cdclk.hw.cdclk);
> >   	if (cdctl == expected)
> >   		/* All well; nothing to sanitize */
> > @@ -1748,10 +1748,10 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
> >   	/* force cdclk programming */
> > -	dev_priv->cdclk.hw.cdclk = 0;
> > +	i915->cdclk.hw.cdclk = 0;
> >   	/* force full PLL disable + enable */
> > -	dev_priv->cdclk.hw.vco = -1;
> > +	i915->cdclk.hw.vco = -1;
> >   }
> >   static int icl_calc_cdclk(int min_cdclk, unsigned int ref)
> > @@ -1781,11 +1781,11 @@ static int icl_calc_cdclk(int min_cdclk, unsigned int ref)
> >   		return ranges[0];
> >   }
> > -static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> > +static int icl_calc_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk)
> >   {
> >   	int ratio;
> > -	if (cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (cdclk == i915->cdclk.hw.bypass)
> >   		return 0;
> >   	switch (cdclk) {
> > @@ -1795,21 +1795,21 @@ static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk)
> >   	case 307200:
> >   	case 556800:
> >   	case 652800:
> > -		WARN_ON(dev_priv->cdclk.hw.ref != 19200 &&
> > -			dev_priv->cdclk.hw.ref != 38400);
> > +		WARN_ON(i915->cdclk.hw.ref != 19200 &&
> > +			i915->cdclk.hw.ref != 38400);
> >   		break;
> >   	case 312000:
> >   	case 552000:
> >   	case 648000:
> > -		WARN_ON(dev_priv->cdclk.hw.ref != 24000);
> > +		WARN_ON(i915->cdclk.hw.ref != 24000);
> >   	}
> > -	ratio = cdclk / (dev_priv->cdclk.hw.ref / 2);
> > +	ratio = cdclk / (i915->cdclk.hw.ref / 2);
> > -	return dev_priv->cdclk.hw.ref * ratio;
> > +	return i915->cdclk.hw.ref * ratio;
> >   }
> > -static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void icl_set_cdclk(struct drm_i915_private *i915,
> >   			  const struct intel_cdclk_state *cdclk_state,
> >   			  enum pipe pipe)
> >   {
> > @@ -1817,7 +1817,7 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	unsigned int vco = cdclk_state->vco;
> >   	int ret;
> > -	ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				SKL_CDCLK_PREPARE_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE,
> >   				SKL_CDCLK_READY_FOR_CHANGE, 3);
> > @@ -1827,12 +1827,12 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	if (dev_priv->cdclk.hw.vco != 0 &&
> > -	    dev_priv->cdclk.hw.vco != vco)
> > -		cnl_cdclk_pll_disable(dev_priv);
> > +	if (i915->cdclk.hw.vco != 0 &&
> > +	    i915->cdclk.hw.vco != vco)
> > +		cnl_cdclk_pll_disable(i915);
> > -	if (dev_priv->cdclk.hw.vco != vco)
> > -		cnl_cdclk_pll_enable(dev_priv, vco);
> > +	if (i915->cdclk.hw.vco != vco)
> > +		cnl_cdclk_pll_enable(i915, vco);
> >   	/*
> >   	 * On ICL CD2X_DIV can only be 1, so we'll never end up changing the
> > @@ -1842,16 +1842,16 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CDCLK_CTL, ICL_CDCLK_CD2X_PIPE_NONE |
> >   			      skl_cdclk_decimal(cdclk));
> > -	sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL,
> > +	sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL,
> >   				cdclk_state->voltage_level);
> > -	intel_update_cdclk(dev_priv);
> > +	intel_update_cdclk(i915);
> >   	/*
> >   	 * Can't read out the voltage level :(
> >   	 * Let's just assume everything is as expected.
> >   	 */
> > -	dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> > +	i915->cdclk.hw.voltage_level = cdclk_state->voltage_level;
> >   }
> >   static u8 icl_calc_voltage_level(int cdclk)
> > @@ -1864,7 +1864,7 @@ static u8 icl_calc_voltage_level(int cdclk)
> >   		return 0;
> >   }
> > -static void icl_get_cdclk(struct drm_i915_private *dev_priv,
> > +static void icl_get_cdclk(struct drm_i915_private *i915,
> >   			  struct intel_cdclk_state *cdclk_state)
> >   {
> >   	u32 val;
> > @@ -1915,17 +1915,17 @@ static void icl_get_cdclk(struct drm_i915_private *dev_priv,
> >   		icl_calc_voltage_level(cdclk_state->cdclk);
> >   }
> > -static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> > +static void icl_init_cdclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_cdclk_state sanitized_state;
> >   	u32 val;
> > -	/* This sets dev_priv->cdclk.hw. */
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > +	/* This sets i915->cdclk.hw. */
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> >   	/* This means CDCLK disabled. */
> > -	if (dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass)
> > +	if (i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass)
> >   		goto sanitize;
> >   	val = I915_READ(CDCLK_CTL);
> > @@ -1934,7 +1934,7 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> >   		goto sanitize;
> >   	if ((val & CDCLK_FREQ_DECIMAL_MASK) !=
> > -	    skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk))
> > +	    skl_cdclk_decimal(i915->cdclk.hw.cdclk))
> >   		goto sanitize;
> >   	return;
> > @@ -1942,62 +1942,62 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv)
> >   sanitize:
> >   	DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n");
> > -	sanitized_state.ref = dev_priv->cdclk.hw.ref;
> > +	sanitized_state.ref = i915->cdclk.hw.ref;
> >   	sanitized_state.cdclk = icl_calc_cdclk(0, sanitized_state.ref);
> > -	sanitized_state.vco = icl_calc_cdclk_pll_vco(dev_priv,
> > +	sanitized_state.vco = icl_calc_cdclk_pll_vco(i915,
> >   						     sanitized_state.cdclk);
> >   	sanitized_state.voltage_level =
> >   				icl_calc_voltage_level(sanitized_state.cdclk);
> > -	icl_set_cdclk(dev_priv, &sanitized_state, INVALID_PIPE);
> > +	icl_set_cdclk(i915, &sanitized_state, INVALID_PIPE);
> >   }
> > -static void icl_uninit_cdclk(struct drm_i915_private *dev_priv)
> > +static void icl_uninit_cdclk(struct drm_i915_private *i915)
> >   {
> > -	struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> > +	struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
> >   	cdclk_state.cdclk = cdclk_state.bypass;
> >   	cdclk_state.vco = 0;
> >   	cdclk_state.voltage_level = icl_calc_voltage_level(cdclk_state.cdclk);
> > -	icl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	icl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> > -static void cnl_init_cdclk(struct drm_i915_private *dev_priv)
> > +static void cnl_init_cdclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_cdclk_state cdclk_state;
> > -	cnl_sanitize_cdclk(dev_priv);
> > +	cnl_sanitize_cdclk(i915);
> > -	if (dev_priv->cdclk.hw.cdclk != 0 &&
> > -	    dev_priv->cdclk.hw.vco != 0)
> > +	if (i915->cdclk.hw.cdclk != 0 &&
> > +	    i915->cdclk.hw.vco != 0)
> >   		return;
> > -	cdclk_state = dev_priv->cdclk.hw;
> > +	cdclk_state = i915->cdclk.hw;
> >   	cdclk_state.cdclk = cnl_calc_cdclk(0);
> > -	cdclk_state.vco = cnl_cdclk_pll_vco(dev_priv, cdclk_state.cdclk);
> > +	cdclk_state.vco = cnl_cdclk_pll_vco(i915, cdclk_state.cdclk);
> >   	cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk);
> > -	cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> > -static void cnl_uninit_cdclk(struct drm_i915_private *dev_priv)
> > +static void cnl_uninit_cdclk(struct drm_i915_private *i915)
> >   {
> > -	struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw;
> > +	struct intel_cdclk_state cdclk_state = i915->cdclk.hw;
> >   	cdclk_state.cdclk = cdclk_state.bypass;
> >   	cdclk_state.vco = 0;
> >   	cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk);
> > -	cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE);
> > +	cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE);
> >   }
> >   /**
> >    * intel_cdclk_init - Initialize CDCLK
> >    * @i915: i915 device
> >    *
> > - * Initialize CDCLK. This consists mainly of initializing dev_priv->cdclk.hw and
> > + * Initialize CDCLK. This consists mainly of initializing i915->cdclk.hw and
> >    * sanitizing the state of the hardware if needed. This is generally done only
> >    * during the display core initialization sequence, after which the DMC will
> >    * take care of turning CDCLK off/on as needed.
> > @@ -2051,19 +2051,19 @@ bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a,
> >   /**
> >    * intel_cdclk_needs_cd2x_update - Determine if two CDCLK states require a cd2x divider update
> > - * @dev_priv: Not a CDCLK state, it's the drm_i915_private!
> > + * @i915: Not a CDCLK state, it's the drm_i915_private!
> >    * @a: first CDCLK state
> >    * @b: second CDCLK state
> >    *
> >    * Returns:
> >    * True if the CDCLK states require just a cd2x divider update, false if not.
> >    */
> > -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv,
> > +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915,
> >   				   const struct intel_cdclk_state *a,
> >   				   const struct intel_cdclk_state *b)
> >   {
> >   	/* Older hw doesn't have the capability */
> > -	if (INTEL_GEN(dev_priv) < 10 && !IS_GEN9_LP(dev_priv))
> > +	if (INTEL_GEN(i915) < 10 && !IS_GEN9_LP(i915))
> >   		return false;
> >   	return a->cdclk != b->cdclk &&
> > @@ -2100,10 +2100,10 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a,
> >    */
> >   void intel_cdclk_swap_state(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> > -	swap(state->cdclk.logical, dev_priv->cdclk.logical);
> > -	swap(state->cdclk.actual, dev_priv->cdclk.actual);
> > +	swap(state->cdclk.logical, i915->cdclk.logical);
> > +	swap(state->cdclk.actual, i915->cdclk.actual);
> >   }
> >   void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state,
> > @@ -2117,37 +2117,37 @@ void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state,
> >   /**
> >    * intel_set_cdclk - Push the CDCLK state to the hardware
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @cdclk_state: new CDCLK state
> >    * @pipe: pipe with which to synchronize the update
> >    *
> >    * Program the hardware based on the passed in CDCLK state,
> >    * if necessary.
> >    */
> > -static void intel_set_cdclk(struct drm_i915_private *dev_priv,
> > +static void intel_set_cdclk(struct drm_i915_private *i915,
> >   			    const struct intel_cdclk_state *cdclk_state,
> >   			    enum pipe pipe)
> >   {
> > -	if (!intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state))
> > +	if (!intel_cdclk_changed(&i915->cdclk.hw, cdclk_state))
> >   		return;
> > -	if (WARN_ON_ONCE(!dev_priv->display.set_cdclk))
> > +	if (WARN_ON_ONCE(!i915->display.set_cdclk))
> >   		return;
> >   	intel_dump_cdclk_state(cdclk_state, "Changing CDCLK to");
> > -	dev_priv->display.set_cdclk(dev_priv, cdclk_state, pipe);
> > +	i915->display.set_cdclk(i915, cdclk_state, pipe);
> > -	if (WARN(intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state),
> > +	if (WARN(intel_cdclk_changed(&i915->cdclk.hw, cdclk_state),
> >   		 "cdclk state doesn't match!\n")) {
> > -		intel_dump_cdclk_state(&dev_priv->cdclk.hw, "[hw state]");
> > +		intel_dump_cdclk_state(&i915->cdclk.hw, "[hw state]");
> >   		intel_dump_cdclk_state(cdclk_state, "[sw state]");
> >   	}
> >   }
> >   /**
> >    * intel_set_cdclk_pre_plane_update - Push the CDCLK state to the hardware
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @old_state: old CDCLK state
> >    * @new_state: new CDCLK state
> >    * @pipe: pipe with which to synchronize the update
> > @@ -2156,18 +2156,18 @@ static void intel_set_cdclk(struct drm_i915_private *dev_priv,
> >    * in CDCLK state, if necessary.
> >    */
> >   void
> > -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> > +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915,
> >   				 const struct intel_cdclk_state *old_state,
> >   				 const struct intel_cdclk_state *new_state,
> >   				 enum pipe pipe)
> >   {
> >   	if (pipe == INVALID_PIPE || old_state->cdclk <= new_state->cdclk)
> > -		intel_set_cdclk(dev_priv, new_state, pipe);
> > +		intel_set_cdclk(i915, new_state, pipe);
> >   }
> >   /**
> >    * intel_set_cdclk_post_plane_update - Push the CDCLK state to the hardware
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @old_state: old CDCLK state
> >    * @new_state: new CDCLK state
> >    * @pipe: pipe with which to synchronize the update
> > @@ -2176,24 +2176,24 @@ intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> >    * in CDCLK state, if necessary.
> >    */
> >   void
> > -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv,
> > +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915,
> >   				  const struct intel_cdclk_state *old_state,
> >   				  const struct intel_cdclk_state *new_state,
> >   				  enum pipe pipe)
> >   {
> >   	if (pipe != INVALID_PIPE && old_state->cdclk > new_state->cdclk)
> > -		intel_set_cdclk(dev_priv, new_state, pipe);
> > +		intel_set_cdclk(i915, new_state, pipe);
> >   }
> > -static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv,
> > +static int intel_pixel_rate_to_cdclk(struct drm_i915_private *i915,
> >   				     int pixel_rate)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		return DIV_ROUND_UP(pixel_rate, 2);
> > -	else if (IS_GEN(dev_priv, 9) ||
> > -		 IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> > +	else if (IS_GEN(i915, 9) ||
> > +		 IS_BROADWELL(i915) || IS_HASWELL(i915))
> >   		return pixel_rate;
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915))
> >   		return DIV_ROUND_UP(pixel_rate * 100, 95);
> >   	else
> >   		return DIV_ROUND_UP(pixel_rate * 100, 90);
> > @@ -2201,17 +2201,17 @@ static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv,
> >   int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(crtc_state->base.crtc->dev);
> >   	int min_cdclk;
> >   	if (!crtc_state->base.enable)
> >   		return 0;
> > -	min_cdclk = intel_pixel_rate_to_cdclk(dev_priv, crtc_state->pixel_rate);
> > +	min_cdclk = intel_pixel_rate_to_cdclk(i915, crtc_state->pixel_rate);
> >   	/* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */
> > -	if (IS_BROADWELL(dev_priv) && hsw_crtc_state_ips_capable(crtc_state))
> > +	if (IS_BROADWELL(i915) && hsw_crtc_state_ips_capable(crtc_state))
> >   		min_cdclk = DIV_ROUND_UP(min_cdclk * 100, 95);
> >   	/* BSpec says "Do not use DisplayPort with CDCLK less than 432 MHz,
> > @@ -2223,10 +2223,10 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   	    crtc_state->has_audio &&
> >   	    crtc_state->port_clock >= 540000 &&
> >   	    crtc_state->lane_count == 4) {
> > -		if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) {
> > +		if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) {
> >   			/* Display WA #1145: glk,cnl */
> >   			min_cdclk = max(316800, min_cdclk);
> > -		} else if (IS_GEN(dev_priv, 9) || IS_BROADWELL(dev_priv)) {
> > +		} else if (IS_GEN(i915, 9) || IS_BROADWELL(i915)) {
> >   			/* Display WA #1144: skl,bxt */
> >   			min_cdclk = max(432000, min_cdclk);
> >   		}
> > @@ -2236,7 +2236,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   	 * According to BSpec, "The CD clock frequency must be at least twice
> >   	 * the frequency of the Azalia BCLK." and BCLK is 96 MHz by default.
> >   	 */
> > -	if (crtc_state->has_audio && INTEL_GEN(dev_priv) >= 9)
> > +	if (crtc_state->has_audio && INTEL_GEN(i915) >= 9)
> >   		min_cdclk = max(2 * 96000, min_cdclk);
> >   	/*
> > @@ -2244,7 +2244,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   	 * than 320000KHz.
> >   	 */
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) &&
> > -	    IS_VALLEYVIEW(dev_priv))
> > +	    IS_VALLEYVIEW(i915))
> >   		min_cdclk = max(320000, min_cdclk);
> >   	/*
> > @@ -2253,12 +2253,12 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   	 * correct for DSI PLL and DE PLL.
> >   	 */
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) &&
> > -	    IS_GEMINILAKE(dev_priv))
> > +	    IS_GEMINILAKE(i915))
> >   		min_cdclk = max(158400, min_cdclk);
> > -	if (min_cdclk > dev_priv->max_cdclk_freq) {
> > +	if (min_cdclk > i915->max_cdclk_freq) {
> >   		DRM_DEBUG_KMS("required cdclk (%d kHz) exceeds max (%d kHz)\n",
> > -			      min_cdclk, dev_priv->max_cdclk_freq);
> > +			      min_cdclk, i915->max_cdclk_freq);
> >   		return -EINVAL;
> >   	}
> > @@ -2267,13 +2267,13 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state)
> >   static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc *crtc;
> >   	struct intel_crtc_state *crtc_state;
> >   	int min_cdclk, i;
> >   	enum pipe pipe;
> > -	memcpy(state->min_cdclk, dev_priv->min_cdclk,
> > +	memcpy(state->min_cdclk, i915->min_cdclk,
> >   	       sizeof(state->min_cdclk));
> >   	for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> > @@ -2285,7 +2285,7 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> >   	}
> >   	min_cdclk = state->cdclk.force_min_cdclk;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		min_cdclk = max(state->min_cdclk[pipe], min_cdclk);
> >   	return min_cdclk;
> > @@ -2302,14 +2302,14 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state)
> >    */
> >   static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc *crtc;
> >   	struct intel_crtc_state *crtc_state;
> >   	u8 min_voltage_level;
> >   	int i;
> >   	enum pipe pipe;
> > -	memcpy(state->min_voltage_level, dev_priv->min_voltage_level,
> > +	memcpy(state->min_voltage_level, i915->min_voltage_level,
> >   	       sizeof(state->min_voltage_level));
> >   	for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> > @@ -2321,7 +2321,7 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
> >   	}
> >   	min_voltage_level = 0;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		min_voltage_level = max(state->min_voltage_level[pipe],
> >   					min_voltage_level);
> > @@ -2330,25 +2330,25 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state)
> >   static int vlv_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	int min_cdclk, cdclk;
> >   	min_cdclk = intel_compute_min_cdclk(state);
> >   	if (min_cdclk < 0)
> >   		return min_cdclk;
> > -	cdclk = vlv_calc_cdclk(dev_priv, min_cdclk);
> > +	cdclk = vlv_calc_cdclk(i915, min_cdclk);
> >   	state->cdclk.logical.cdclk = cdclk;
> >   	state->cdclk.logical.voltage_level =
> > -		vlv_calc_voltage_level(dev_priv, cdclk);
> > +		vlv_calc_voltage_level(i915, cdclk);
> >   	if (!state->active_crtcs) {
> > -		cdclk = vlv_calc_cdclk(dev_priv, state->cdclk.force_min_cdclk);
> > +		cdclk = vlv_calc_cdclk(i915, state->cdclk.force_min_cdclk);
> >   		state->cdclk.actual.cdclk = cdclk;
> >   		state->cdclk.actual.voltage_level =
> > -			vlv_calc_voltage_level(dev_priv, cdclk);
> > +			vlv_calc_voltage_level(i915, cdclk);
> >   	} else {
> >   		state->cdclk.actual = state->cdclk.logical;
> >   	}
> > @@ -2389,14 +2389,14 @@ static int bdw_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   static int skl_dpll0_vco(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc *crtc;
> >   	struct intel_crtc_state *crtc_state;
> >   	int vco, i;
> >   	vco = state->cdclk.logical.vco;
> >   	if (!vco)
> > -		vco = dev_priv->skl_preferred_vco_freq;
> > +		vco = i915->skl_preferred_vco_freq;
> >   	for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
> >   		if (!crtc_state->base.enable)
> > @@ -2460,19 +2460,19 @@ static int skl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	int min_cdclk, cdclk, vco;
> >   	min_cdclk = intel_compute_min_cdclk(state);
> >   	if (min_cdclk < 0)
> >   		return min_cdclk;
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		cdclk = glk_calc_cdclk(min_cdclk);
> > -		vco = glk_de_pll_vco(dev_priv, cdclk);
> > +		vco = glk_de_pll_vco(i915, cdclk);
> >   	} else {
> >   		cdclk = bxt_calc_cdclk(min_cdclk);
> > -		vco = bxt_de_pll_vco(dev_priv, cdclk);
> > +		vco = bxt_de_pll_vco(i915, cdclk);
> >   	}
> >   	state->cdclk.logical.vco = vco;
> > @@ -2481,12 +2481,12 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   		bxt_calc_voltage_level(cdclk);
> >   	if (!state->active_crtcs) {
> > -		if (IS_GEMINILAKE(dev_priv)) {
> > +		if (IS_GEMINILAKE(i915)) {
> >   			cdclk = glk_calc_cdclk(state->cdclk.force_min_cdclk);
> > -			vco = glk_de_pll_vco(dev_priv, cdclk);
> > +			vco = glk_de_pll_vco(i915, cdclk);
> >   		} else {
> >   			cdclk = bxt_calc_cdclk(state->cdclk.force_min_cdclk);
> > -			vco = bxt_de_pll_vco(dev_priv, cdclk);
> > +			vco = bxt_de_pll_vco(i915, cdclk);
> >   		}
> >   		state->cdclk.actual.vco = vco;
> > @@ -2502,7 +2502,7 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	int min_cdclk, cdclk, vco;
> >   	min_cdclk = intel_compute_min_cdclk(state);
> > @@ -2510,7 +2510,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   		return min_cdclk;
> >   	cdclk = cnl_calc_cdclk(min_cdclk);
> > -	vco = cnl_cdclk_pll_vco(dev_priv, cdclk);
> > +	vco = cnl_cdclk_pll_vco(i915, cdclk);
> >   	state->cdclk.logical.vco = vco;
> >   	state->cdclk.logical.cdclk = cdclk;
> > @@ -2520,7 +2520,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   	if (!state->active_crtcs) {
> >   		cdclk = cnl_calc_cdclk(state->cdclk.force_min_cdclk);
> > -		vco = cnl_cdclk_pll_vco(dev_priv, cdclk);
> > +		vco = cnl_cdclk_pll_vco(i915, cdclk);
> >   		state->cdclk.actual.vco = vco;
> >   		state->cdclk.actual.cdclk = cdclk;
> > @@ -2535,7 +2535,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	unsigned int ref = state->cdclk.logical.ref;
> >   	int min_cdclk, cdclk, vco;
> > @@ -2544,7 +2544,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   		return min_cdclk;
> >   	cdclk = icl_calc_cdclk(min_cdclk, ref);
> > -	vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk);
> > +	vco = icl_calc_cdclk_pll_vco(i915, cdclk);
> >   	state->cdclk.logical.vco = vco;
> >   	state->cdclk.logical.cdclk = cdclk;
> > @@ -2554,7 +2554,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   	if (!state->active_crtcs) {
> >   		cdclk = icl_calc_cdclk(state->cdclk.force_min_cdclk, ref);
> > -		vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk);
> > +		vco = icl_calc_cdclk_pll_vco(i915, cdclk);
> >   		state->cdclk.actual.vco = vco;
> >   		state->cdclk.actual.cdclk = cdclk;
> > @@ -2567,18 +2567,18 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state)
> >   	return 0;
> >   }
> > -static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv)
> > +static int intel_compute_max_dotclk(struct drm_i915_private *i915)
> >   {
> > -	int max_cdclk_freq = dev_priv->max_cdclk_freq;
> > +	int max_cdclk_freq = i915->max_cdclk_freq;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		return 2 * max_cdclk_freq;
> > -	else if (IS_GEN(dev_priv, 9) ||
> > -		 IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> > +	else if (IS_GEN(i915, 9) ||
> > +		 IS_BROADWELL(i915) || IS_HASWELL(i915))
> >   		return max_cdclk_freq;
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915))
> >   		return max_cdclk_freq*95/100;
> > -	else if (INTEL_GEN(dev_priv) < 4)
> > +	else if (INTEL_GEN(i915) < 4)
> >   		return 2*max_cdclk_freq*90/100;
> >   	else
> >   		return max_cdclk_freq*90/100;
> > @@ -2586,26 +2586,26 @@ static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_update_max_cdclk - Determine the maximum support CDCLK frequency
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    *
> >    * Determine the maximum CDCLK frequency the platform supports, and also
> >    * derive the maximum dot clock frequency the maximum CDCLK frequency
> >    * allows.
> >    */
> > -void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> > +void intel_update_max_cdclk(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		if (dev_priv->cdclk.hw.ref == 24000)
> > -			dev_priv->max_cdclk_freq = 648000;
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		if (i915->cdclk.hw.ref == 24000)
> > +			i915->max_cdclk_freq = 648000;
> >   		else
> > -			dev_priv->max_cdclk_freq = 652800;
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > -		dev_priv->max_cdclk_freq = 528000;
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > +			i915->max_cdclk_freq = 652800;
> > +	} else if (IS_CANNONLAKE(i915)) {
> > +		i915->max_cdclk_freq = 528000;
> > +	} else if (IS_GEN9_BC(i915)) {
> >   		u32 limit = I915_READ(SKL_DFSM) & SKL_DFSM_CDCLK_LIMIT_MASK;
> >   		int max_cdclk, vco;
> > -		vco = dev_priv->skl_preferred_vco_freq;
> > +		vco = i915->skl_preferred_vco_freq;
> >   		WARN_ON(vco != 8100000 && vco != 8640000);
> >   		/*
> > @@ -2622,12 +2622,12 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> >   		else
> >   			max_cdclk = 308571;
> > -		dev_priv->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco);
> > -	} else if (IS_GEMINILAKE(dev_priv)) {
> > -		dev_priv->max_cdclk_freq = 316800;
> > -	} else if (IS_BROXTON(dev_priv)) {
> > -		dev_priv->max_cdclk_freq = 624000;
> > -	} else if (IS_BROADWELL(dev_priv))  {
> > +		i915->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco);
> > +	} else if (IS_GEMINILAKE(i915)) {
> > +		i915->max_cdclk_freq = 316800;
> > +	} else if (IS_BROXTON(i915)) {
> > +		i915->max_cdclk_freq = 624000;
> > +	} else if (IS_BROADWELL(i915))  {
> >   		/*
> >   		 * FIXME with extra cooling we can allow
> >   		 * 540 MHz for ULX and 675 Mhz for ULT.
> > @@ -2635,40 +2635,40 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv)
> >   		 * available? PCI ID, VTB, something else?
> >   		 */
> >   		if (I915_READ(FUSE_STRAP) & HSW_CDCLK_LIMIT)
> > -			dev_priv->max_cdclk_freq = 450000;
> > -		else if (IS_BDW_ULX(dev_priv))
> > -			dev_priv->max_cdclk_freq = 450000;
> > -		else if (IS_BDW_ULT(dev_priv))
> > -			dev_priv->max_cdclk_freq = 540000;
> > +			i915->max_cdclk_freq = 450000;
> > +		else if (IS_BDW_ULX(i915))
> > +			i915->max_cdclk_freq = 450000;
> > +		else if (IS_BDW_ULT(i915))
> > +			i915->max_cdclk_freq = 540000;
> >   		else
> > -			dev_priv->max_cdclk_freq = 675000;
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > -		dev_priv->max_cdclk_freq = 320000;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > -		dev_priv->max_cdclk_freq = 400000;
> > +			i915->max_cdclk_freq = 675000;
> > +	} else if (IS_CHERRYVIEW(i915)) {
> > +		i915->max_cdclk_freq = 320000;
> > +	} else if (IS_VALLEYVIEW(i915)) {
> > +		i915->max_cdclk_freq = 400000;
> >   	} else {
> >   		/* otherwise assume cdclk is fixed */
> > -		dev_priv->max_cdclk_freq = dev_priv->cdclk.hw.cdclk;
> > +		i915->max_cdclk_freq = i915->cdclk.hw.cdclk;
> >   	}
> > -	dev_priv->max_dotclk_freq = intel_compute_max_dotclk(dev_priv);
> > +	i915->max_dotclk_freq = intel_compute_max_dotclk(i915);
> >   	DRM_DEBUG_DRIVER("Max CD clock rate: %d kHz\n",
> > -			 dev_priv->max_cdclk_freq);
> > +			 i915->max_cdclk_freq);
> >   	DRM_DEBUG_DRIVER("Max dotclock rate: %d kHz\n",
> > -			 dev_priv->max_dotclk_freq);
> > +			 i915->max_dotclk_freq);
> >   }
> >   /**
> >    * intel_update_cdclk - Determine the current CDCLK frequency
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    *
> >    * Determine the current CDCLK frequency.
> >    */
> > -void intel_update_cdclk(struct drm_i915_private *dev_priv)
> > +void intel_update_cdclk(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->display.get_cdclk(dev_priv, &dev_priv->cdclk.hw);
> > +	i915->display.get_cdclk(i915, &i915->cdclk.hw);
> >   	/*
> >   	 * 9:0 CMBUS [sic] CDCLK frequency (cdfreq):
> > @@ -2676,12 +2676,12 @@ void intel_update_cdclk(struct drm_i915_private *dev_priv)
> >   	 * of cdclk that generates 4MHz reference clock freq which is used to
> >   	 * generate GMBus clock. This will vary with the cdclk freq.
> >   	 */
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		I915_WRITE(GMBUSFREQ_VLV,
> > -			   DIV_ROUND_UP(dev_priv->cdclk.hw.cdclk, 1000));
> > +			   DIV_ROUND_UP(i915->cdclk.hw.cdclk, 1000));
> >   }
> > -static int cnp_rawclk(struct drm_i915_private *dev_priv)
> > +static int cnp_rawclk(struct drm_i915_private *i915)
> >   {
> >   	u32 rawclk;
> >   	int divider, fraction;
> > @@ -2702,7 +2702,7 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv)
> >   		rawclk |= CNP_RAWCLK_DEN(DIV_ROUND_CLOSEST(numerator * 1000,
> >   							   fraction) - 1);
> > -		if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP)
> > +		if (INTEL_PCH_TYPE(i915) >= PCH_ICP)
> >   			rawclk |= ICP_RAWCLK_NUM(numerator);
> >   	}
> > @@ -2710,19 +2710,19 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv)
> >   	return divider + fraction;
> >   }
> > -static int pch_rawclk(struct drm_i915_private *dev_priv)
> > +static int pch_rawclk(struct drm_i915_private *i915)
> >   {
> >   	return (I915_READ(PCH_RAWCLK_FREQ) & RAWCLK_FREQ_MASK) * 1000;
> >   }
> > -static int vlv_hrawclk(struct drm_i915_private *dev_priv)
> > +static int vlv_hrawclk(struct drm_i915_private *i915)
> >   {
> >   	/* RAWCLK_FREQ_VLV register updated from power well code */
> > -	return vlv_get_cck_clock_hpll(dev_priv, "hrawclk",
> > +	return vlv_get_cck_clock_hpll(i915, "hrawclk",
> >   				      CCK_DISPLAY_REF_CLOCK_CONTROL);
> >   }
> > -static int g4x_hrawclk(struct drm_i915_private *dev_priv)
> > +static int g4x_hrawclk(struct drm_i915_private *i915)
> >   {
> >   	u32 clkcfg;
> > @@ -2750,104 +2750,104 @@ static int g4x_hrawclk(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_update_rawclk - Determine the current RAWCLK frequency
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    *
> >    * Determine the current RAWCLK frequency. RAWCLK is a fixed
> >    * frequency clock so this needs to done only once.
> >    */
> > -void intel_update_rawclk(struct drm_i915_private *dev_priv)
> > -{
> > -	if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP)
> > -		dev_priv->rawclk_freq = cnp_rawclk(dev_priv);
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > -		dev_priv->rawclk_freq = pch_rawclk(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->rawclk_freq = vlv_hrawclk(dev_priv);
> > -	else if (IS_G4X(dev_priv) || IS_PINEVIEW(dev_priv))
> > -		dev_priv->rawclk_freq = g4x_hrawclk(dev_priv);
> > +void intel_update_rawclk(struct drm_i915_private *i915)
> > +{
> > +	if (INTEL_PCH_TYPE(i915) >= PCH_CNP)
> > +		i915->rawclk_freq = cnp_rawclk(i915);
> > +	else if (HAS_PCH_SPLIT(i915))
> > +		i915->rawclk_freq = pch_rawclk(i915);
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->rawclk_freq = vlv_hrawclk(i915);
> > +	else if (IS_G4X(i915) || IS_PINEVIEW(i915))
> > +		i915->rawclk_freq = g4x_hrawclk(i915);
> >   	else
> >   		/* no rawclk on other platforms, or no need to know it */
> >   		return;
> > -	DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", dev_priv->rawclk_freq);
> > +	DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", i915->rawclk_freq);
> >   }
> >   /**
> >    * intel_init_cdclk_hooks - Initialize CDCLK related modesetting hooks
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    */
> > -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv)
> > -{
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		dev_priv->display.set_cdclk = icl_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = icl_modeset_calc_cdclk;
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > -		dev_priv->display.set_cdclk = cnl_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk;
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > -		dev_priv->display.set_cdclk = bxt_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk;
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > -		dev_priv->display.set_cdclk = skl_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = skl_modeset_calc_cdclk;
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		dev_priv->display.set_cdclk = bdw_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk;
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > -		dev_priv->display.set_cdclk = chv_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > -		dev_priv->display.set_cdclk = vlv_set_cdclk;
> > -		dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> > +void intel_init_cdclk_hooks(struct drm_i915_private *i915)
> > +{
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		i915->display.set_cdclk = icl_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = icl_modeset_calc_cdclk;
> > +	} else if (IS_CANNONLAKE(i915)) {
> > +		i915->display.set_cdclk = cnl_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk;
> > +	} else if (IS_GEN9_LP(i915)) {
> > +		i915->display.set_cdclk = bxt_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk;
> > +	} else if (IS_GEN9_BC(i915)) {
> > +		i915->display.set_cdclk = skl_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = skl_modeset_calc_cdclk;
> > +	} else if (IS_BROADWELL(i915)) {
> > +		i915->display.set_cdclk = bdw_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk;
> > +	} else if (IS_CHERRYVIEW(i915)) {
> > +		i915->display.set_cdclk = chv_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> > +	} else if (IS_VALLEYVIEW(i915)) {
> > +		i915->display.set_cdclk = vlv_set_cdclk;
> > +		i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		dev_priv->display.get_cdclk = icl_get_cdclk;
> > -	else if (IS_CANNONLAKE(dev_priv))
> > -		dev_priv->display.get_cdclk = cnl_get_cdclk;
> > -	else if (IS_GEN9_LP(dev_priv))
> > -		dev_priv->display.get_cdclk = bxt_get_cdclk;
> > -	else if (IS_GEN9_BC(dev_priv))
> > -		dev_priv->display.get_cdclk = skl_get_cdclk;
> > -	else if (IS_BROADWELL(dev_priv))
> > -		dev_priv->display.get_cdclk = bdw_get_cdclk;
> > -	else if (IS_HASWELL(dev_priv))
> > -		dev_priv->display.get_cdclk = hsw_get_cdclk;
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->display.get_cdclk = vlv_get_cdclk;
> > -	else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> > -	else if (IS_GEN(dev_priv, 5))
> > -		dev_priv->display.get_cdclk = fixed_450mhz_get_cdclk;
> > -	else if (IS_GM45(dev_priv))
> > -		dev_priv->display.get_cdclk = gm45_get_cdclk;
> > -	else if (IS_G45(dev_priv))
> > -		dev_priv->display.get_cdclk = g33_get_cdclk;
> > -	else if (IS_I965GM(dev_priv))
> > -		dev_priv->display.get_cdclk = i965gm_get_cdclk;
> > -	else if (IS_I965G(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> > -	else if (IS_PINEVIEW(dev_priv))
> > -		dev_priv->display.get_cdclk = pnv_get_cdclk;
> > -	else if (IS_G33(dev_priv))
> > -		dev_priv->display.get_cdclk = g33_get_cdclk;
> > -	else if (IS_I945GM(dev_priv))
> > -		dev_priv->display.get_cdclk = i945gm_get_cdclk;
> > -	else if (IS_I945G(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk;
> > -	else if (IS_I915GM(dev_priv))
> > -		dev_priv->display.get_cdclk = i915gm_get_cdclk;
> > -	else if (IS_I915G(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_333mhz_get_cdclk;
> > -	else if (IS_I865G(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_266mhz_get_cdclk;
> > -	else if (IS_I85X(dev_priv))
> > -		dev_priv->display.get_cdclk = i85x_get_cdclk;
> > -	else if (IS_I845G(dev_priv))
> > -		dev_priv->display.get_cdclk = fixed_200mhz_get_cdclk;
> > +	if (INTEL_GEN(i915) >= 11)
> > +		i915->display.get_cdclk = icl_get_cdclk;
> > +	else if (IS_CANNONLAKE(i915))
> > +		i915->display.get_cdclk = cnl_get_cdclk;
> > +	else if (IS_GEN9_LP(i915))
> > +		i915->display.get_cdclk = bxt_get_cdclk;
> > +	else if (IS_GEN9_BC(i915))
> > +		i915->display.get_cdclk = skl_get_cdclk;
> > +	else if (IS_BROADWELL(i915))
> > +		i915->display.get_cdclk = bdw_get_cdclk;
> > +	else if (IS_HASWELL(i915))
> > +		i915->display.get_cdclk = hsw_get_cdclk;
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->display.get_cdclk = vlv_get_cdclk;
> > +	else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> > +		i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> > +	else if (IS_GEN(i915, 5))
> > +		i915->display.get_cdclk = fixed_450mhz_get_cdclk;
> > +	else if (IS_GM45(i915))
> > +		i915->display.get_cdclk = gm45_get_cdclk;
> > +	else if (IS_G45(i915))
> > +		i915->display.get_cdclk = g33_get_cdclk;
> > +	else if (IS_I965GM(i915))
> > +		i915->display.get_cdclk = i965gm_get_cdclk;
> > +	else if (IS_I965G(i915))
> > +		i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> > +	else if (IS_PINEVIEW(i915))
> > +		i915->display.get_cdclk = pnv_get_cdclk;
> > +	else if (IS_G33(i915))
> > +		i915->display.get_cdclk = g33_get_cdclk;
> > +	else if (IS_I945GM(i915))
> > +		i915->display.get_cdclk = i945gm_get_cdclk;
> > +	else if (IS_I945G(i915))
> > +		i915->display.get_cdclk = fixed_400mhz_get_cdclk;
> > +	else if (IS_I915GM(i915))
> > +		i915->display.get_cdclk = i915gm_get_cdclk;
> > +	else if (IS_I915G(i915))
> > +		i915->display.get_cdclk = fixed_333mhz_get_cdclk;
> > +	else if (IS_I865G(i915))
> > +		i915->display.get_cdclk = fixed_266mhz_get_cdclk;
> > +	else if (IS_I85X(i915))
> > +		i915->display.get_cdclk = i85x_get_cdclk;
> > +	else if (IS_I845G(i915))
> > +		i915->display.get_cdclk = fixed_200mhz_get_cdclk;
> >   	else { /* 830 */
> > -		WARN(!IS_I830(dev_priv),
> > +		WARN(!IS_I830(i915),
> >   		     "Unknown platform. Assuming 133 MHz CDCLK\n");
> > -		dev_priv->display.get_cdclk = fixed_133mhz_get_cdclk;
> > +		i915->display.get_cdclk = fixed_133mhz_get_cdclk;
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_cdclk.h b/drivers/gpu/drm/i915/intel_cdclk.h
> > index 4d6f7f5f8930..4e17102af66c 100644
> > --- a/drivers/gpu/drm/i915/intel_cdclk.h
> > +++ b/drivers/gpu/drm/i915/intel_cdclk.h
> > @@ -18,11 +18,11 @@ struct intel_crtc_state;
> >   int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state);
> >   void intel_cdclk_init(struct drm_i915_private *i915);
> >   void intel_cdclk_uninit(struct drm_i915_private *i915);
> > -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv);
> > -void intel_update_max_cdclk(struct drm_i915_private *dev_priv);
> > -void intel_update_cdclk(struct drm_i915_private *dev_priv);
> > -void intel_update_rawclk(struct drm_i915_private *dev_priv);
> > -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv,
> > +void intel_init_cdclk_hooks(struct drm_i915_private *i915);
> > +void intel_update_max_cdclk(struct drm_i915_private *i915);
> > +void intel_update_cdclk(struct drm_i915_private *i915);
> > +void intel_update_rawclk(struct drm_i915_private *i915);
> > +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915,
> >   				   const struct intel_cdclk_state *a,
> >   				   const struct intel_cdclk_state *b);
> >   bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a,
> > @@ -31,12 +31,12 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a,
> >   			 const struct intel_cdclk_state *b);
> >   void intel_cdclk_swap_state(struct intel_atomic_state *state);
> >   void
> > -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv,
> > +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915,
> >   				 const struct intel_cdclk_state *old_state,
> >   				 const struct intel_cdclk_state *new_state,
> >   				 enum pipe pipe);
> >   void
> > -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv,
> > +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915,
> >   				  const struct intel_cdclk_state *old_state,
> >   				  const struct intel_cdclk_state *new_state,
> >   				  enum pipe pipe);
> > diff --git a/drivers/gpu/drm/i915/intel_color.c b/drivers/gpu/drm/i915/intel_color.c
> > index 0b8cf3e8c963..efc67e4a0ba5 100644
> > --- a/drivers/gpu/drm/i915/intel_color.c
> > +++ b/drivers/gpu/drm/i915/intel_color.c
> > @@ -137,7 +137,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc,
> >   				const u16 coeff[9],
> >   				const u16 postoff[3])
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	I915_WRITE(PIPE_CSC_PREOFF_HI(pipe), preoff[0]);
> > @@ -153,7 +153,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc,
> >   	I915_WRITE(PIPE_CSC_COEFF_RV_GV(pipe), coeff[6] << 16 | coeff[7]);
> >   	I915_WRITE(PIPE_CSC_COEFF_BV(pipe), coeff[8] << 16);
> > -	if (INTEL_GEN(dev_priv) >= 7) {
> > +	if (INTEL_GEN(i915) >= 7) {
> >   		I915_WRITE(PIPE_CSC_POSTOFF_HI(pipe), postoff[0]);
> >   		I915_WRITE(PIPE_CSC_POSTOFF_ME(pipe), postoff[1]);
> >   		I915_WRITE(PIPE_CSC_POSTOFF_LO(pipe), postoff[2]);
> > @@ -165,7 +165,7 @@ static void icl_update_output_csc(struct intel_crtc *crtc,
> >   				  const u16 coeff[9],
> >   				  const u16 postoff[3])
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	I915_WRITE(PIPE_CSC_OUTPUT_PREOFF_HI(pipe), preoff[0]);
> > @@ -188,15 +188,15 @@ static void icl_update_output_csc(struct intel_crtc *crtc,
> >   static bool ilk_csc_limited_range(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	/*
> >   	 * FIXME if there's a gamma LUT after the CSC, we should
> >   	 * do the range compression using the gamma LUT instead.
> >   	 */
> >   	return crtc_state->limited_color_range &&
> > -		(IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) ||
> > -		 IS_GEN_RANGE(dev_priv, 9, 10));
> > +		(IS_HASWELL(i915) || IS_BROADWELL(i915) ||
> > +		 IS_GEN_RANGE(i915, 9, 10));
> >   }
> >   static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state,
> > @@ -254,7 +254,7 @@ static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state,
> >   static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	bool limited_color_range = ilk_csc_limited_range(crtc_state);
> >   	if (crtc_state->base.ctm) {
> > @@ -280,7 +280,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   		 * LUT is needed but CSC is not we need to load an
> >   		 * identity matrix.
> >   		 */
> > -		WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_GEMINILAKE(dev_priv));
> > +		WARN_ON(!IS_CANNONLAKE(i915) && !IS_GEMINILAKE(i915));
> >   		ilk_update_pipe_csc(crtc, ilk_csc_off_zero,
> >   				    ilk_csc_coeff_identity,
> > @@ -293,7 +293,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	if (crtc_state->base.ctm) {
> >   		u16 coeff[9];
> > @@ -322,7 +322,7 @@ static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   static void cherryview_load_csc_matrix(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	if (crtc_state->base.ctm) {
> > @@ -388,15 +388,15 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state,
> >   				    const struct drm_property_blob *blob)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	int i;
> > -	if (HAS_GMCH(dev_priv)) {
> > +	if (HAS_GMCH(i915)) {
> >   		if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI))
> > -			assert_dsi_pll_enabled(dev_priv);
> > +			assert_dsi_pll_enabled(i915);
> >   		else
> > -			assert_pll_enabled(dev_priv, pipe);
> > +			assert_pll_enabled(i915, pipe);
> >   	}
> >   	if (blob) {
> > @@ -408,7 +408,7 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state,
> >   				(drm_color_lut_extract(lut[i].green, 8) << 8) |
> >   				drm_color_lut_extract(lut[i].blue, 8);
> > -			if (HAS_GMCH(dev_priv))
> > +			if (HAS_GMCH(i915))
> >   				I915_WRITE(PALETTE(pipe, i), word);
> >   			else
> >   				I915_WRITE(LGC_PALETTE(pipe, i), word);
> > @@ -424,7 +424,7 @@ static void i9xx_load_luts(const struct intel_crtc_state *crtc_state)
> >   static void i9xx_color_commit(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > @@ -437,7 +437,7 @@ static void i9xx_color_commit(const struct intel_crtc_state *crtc_state)
> >   static void ilk_color_commit(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > @@ -452,7 +452,7 @@ static void ilk_color_commit(const struct intel_crtc_state *crtc_state)
> >   static void hsw_color_commit(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode);
> > @@ -462,7 +462,7 @@ static void hsw_color_commit(const struct intel_crtc_state *crtc_state)
> >   static void skl_color_commit(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val = 0;
> > @@ -479,7 +479,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state)
> >   	I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_load_csc_matrix(crtc_state);
> >   	else
> >   		ilk_load_csc_matrix(crtc_state);
> > @@ -488,7 +488,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state)
> >   static void i965_load_lut_10p6(struct intel_crtc *crtc,
> >   			       const struct drm_property_blob *blob)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -519,7 +519,7 @@ static void i965_load_luts(const struct intel_crtc_state *crtc_state)
> >   static void ilk_load_lut_10(struct intel_crtc *crtc,
> >   			    const struct drm_property_blob *blob)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -556,7 +556,7 @@ static void ivb_load_lut_10(struct intel_crtc *crtc,
> >   			    const struct drm_property_blob *blob,
> >   			    u32 prec_index)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int hw_lut_size = ivb_lut_10_size(prec_index);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> > @@ -583,7 +583,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc,
> >   			    const struct drm_property_blob *blob,
> >   			    u32 prec_index)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int hw_lut_size = ivb_lut_10_size(prec_index);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> > @@ -609,7 +609,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc,
> >   static void ivb_load_lut_10_max(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* Program the max register to clamp values > 1.0. */
> > @@ -622,7 +622,7 @@ static void ivb_load_lut_10_max(struct intel_crtc *crtc)
> >   	 * ToDo: Extend the ABI to be able to program values
> >   	 * from 3.0 to 7.0
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> >   		I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 0), 1 << 16);
> >   		I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 1), 1 << 16);
> >   		I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 2), 1 << 16);
> > @@ -678,9 +678,9 @@ static void bdw_load_luts(const struct intel_crtc_state *crtc_state)
> >   static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> > -	const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> > +	const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size;
> >   	const struct drm_color_lut *lut = crtc_state->base.degamma_lut->data;
> >   	u32 i;
> > @@ -717,9 +717,9 @@ static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state)
> >   static void glk_load_degamma_lut_linear(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> > -	const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> > +	const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size;
> >   	u32 i;
> >   	/*
> > @@ -798,7 +798,7 @@ static u32 chv_cgm_degamma_udw(const struct drm_color_lut *color)
> >   static void chv_load_cgm_degamma(struct intel_crtc *crtc,
> >   				 const struct drm_property_blob *blob)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -825,7 +825,7 @@ static u32 chv_cgm_gamma_udw(const struct drm_color_lut *color)
> >   static void chv_load_cgm_gamma(struct intel_crtc *crtc,
> >   			       const struct drm_property_blob *blob)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_color_lut *lut = blob->data;
> >   	int i, lut_size = drm_color_lut_size(blob);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -860,37 +860,37 @@ static void chv_load_luts(const struct intel_crtc_state *crtc_state)
> >   void intel_color_load_luts(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	dev_priv->display.load_luts(crtc_state);
> > +	i915->display.load_luts(crtc_state);
> >   }
> >   void intel_color_commit(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	dev_priv->display.color_commit(crtc_state);
> > +	i915->display.color_commit(crtc_state);
> >   }
> >   int intel_color_check(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	return dev_priv->display.color_check(crtc_state);
> > +	return i915->display.color_check(crtc_state);
> >   }
> >   void intel_color_get_config(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	if (dev_priv->display.read_luts)
> > -		dev_priv->display.read_luts(crtc_state);
> > +	if (i915->display.read_luts)
> > +		i915->display.read_luts(crtc_state);
> >   }
> >   static bool need_plane_update(struct intel_plane *plane,
> >   			      const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	/*
> >   	 * On pre-SKL the pipe gamma enable and pipe csc enable for
> > @@ -898,7 +898,7 @@ static bool need_plane_update(struct intel_plane *plane,
> >   	 * We have to reconfigure that even if the plane is inactive.
> >   	 */
> >   	return crtc_state->active_planes & BIT(plane->id) ||
> > -		(INTEL_GEN(dev_priv) < 9 &&
> > +		(INTEL_GEN(i915) < 9 &&
> >   		 plane->id == PLANE_PRIMARY);
> >   }
> > @@ -906,7 +906,7 @@ static int
> >   intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_atomic_state *state =
> >   		to_intel_atomic_state(new_crtc_state->base.state);
> >   	const struct intel_crtc_state *old_crtc_state =
> > @@ -921,7 +921,7 @@ intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state)
> >   	    new_crtc_state->csc_enable == old_crtc_state->csc_enable)
> >   		return 0;
> > -	for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +	for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   		struct intel_plane_state *plane_state;
> >   		if (!need_plane_update(plane, new_crtc_state))
> > @@ -956,7 +956,7 @@ static int check_lut_size(const struct drm_property_blob *lut, int expected)
> >   static int check_luts(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	const struct drm_property_blob *gamma_lut = crtc_state->base.gamma_lut;
> >   	const struct drm_property_blob *degamma_lut = crtc_state->base.degamma_lut;
> >   	int gamma_length, degamma_length;
> > @@ -972,10 +972,10 @@ static int check_luts(const struct intel_crtc_state *crtc_state)
> >   		return -EINVAL;
> >   	}
> > -	degamma_length = INTEL_INFO(dev_priv)->color.degamma_lut_size;
> > -	gamma_length = INTEL_INFO(dev_priv)->color.gamma_lut_size;
> > -	degamma_tests = INTEL_INFO(dev_priv)->color.degamma_lut_tests;
> > -	gamma_tests = INTEL_INFO(dev_priv)->color.gamma_lut_tests;
> > +	degamma_length = INTEL_INFO(i915)->color.degamma_lut_size;
> > +	gamma_length = INTEL_INFO(i915)->color.gamma_lut_size;
> > +	degamma_tests = INTEL_INFO(i915)->color.degamma_lut_tests;
> > +	gamma_tests = INTEL_INFO(i915)->color.gamma_lut_tests;
> >   	if (check_lut_size(degamma_lut, degamma_length) ||
> >   	    check_lut_size(gamma_lut, gamma_length))
> > @@ -1255,56 +1255,56 @@ static int icl_color_check(struct intel_crtc_state *crtc_state)
> >   void intel_color_init(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	bool has_ctm = INTEL_INFO(dev_priv)->color.degamma_lut_size != 0;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	bool has_ctm = INTEL_INFO(i915)->color.degamma_lut_size != 0;
> >   	drm_mode_crtc_set_gamma_size(&crtc->base, 256);
> > -	if (HAS_GMCH(dev_priv)) {
> > -		if (IS_CHERRYVIEW(dev_priv)) {
> > -			dev_priv->display.color_check = chv_color_check;
> > -			dev_priv->display.color_commit = i9xx_color_commit;
> > -			dev_priv->display.load_luts = chv_load_luts;
> > -		} else if (INTEL_GEN(dev_priv) >= 4) {
> > -			dev_priv->display.color_check = i9xx_color_check;
> > -			dev_priv->display.color_commit = i9xx_color_commit;
> > -			dev_priv->display.load_luts = i965_load_luts;
> > +	if (HAS_GMCH(i915)) {
> > +		if (IS_CHERRYVIEW(i915)) {
> > +			i915->display.color_check = chv_color_check;
> > +			i915->display.color_commit = i9xx_color_commit;
> > +			i915->display.load_luts = chv_load_luts;
> > +		} else if (INTEL_GEN(i915) >= 4) {
> > +			i915->display.color_check = i9xx_color_check;
> > +			i915->display.color_commit = i9xx_color_commit;
> > +			i915->display.load_luts = i965_load_luts;
> >   		} else {
> > -			dev_priv->display.color_check = i9xx_color_check;
> > -			dev_priv->display.color_commit = i9xx_color_commit;
> > -			dev_priv->display.load_luts = i9xx_load_luts;
> > +			i915->display.color_check = i9xx_color_check;
> > +			i915->display.color_commit = i9xx_color_commit;
> > +			i915->display.load_luts = i9xx_load_luts;
> >   		}
> >   	} else {
> > -		if (INTEL_GEN(dev_priv) >= 11)
> > -			dev_priv->display.color_check = icl_color_check;
> > -		else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > -			dev_priv->display.color_check = glk_color_check;
> > -		else if (INTEL_GEN(dev_priv) >= 7)
> > -			dev_priv->display.color_check = ivb_color_check;
> > +		if (INTEL_GEN(i915) >= 11)
> > +			i915->display.color_check = icl_color_check;
> > +		else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> > +			i915->display.color_check = glk_color_check;
> > +		else if (INTEL_GEN(i915) >= 7)
> > +			i915->display.color_check = ivb_color_check;
> >   		else
> > -			dev_priv->display.color_check = ilk_color_check;
> > +			i915->display.color_check = ilk_color_check;
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > -			dev_priv->display.color_commit = skl_color_commit;
> > -		else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> > -			dev_priv->display.color_commit = hsw_color_commit;
> > +		if (INTEL_GEN(i915) >= 9)
> > +			i915->display.color_commit = skl_color_commit;
> > +		else if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> > +			i915->display.color_commit = hsw_color_commit;
> >   		else
> > -			dev_priv->display.color_commit = ilk_color_commit;
> > -
> > -		if (INTEL_GEN(dev_priv) >= 11)
> > -			dev_priv->display.load_luts = icl_load_luts;
> > -		else if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> > -			dev_priv->display.load_luts = glk_load_luts;
> > -		else if (INTEL_GEN(dev_priv) >= 8)
> > -			dev_priv->display.load_luts = bdw_load_luts;
> > -		else if (INTEL_GEN(dev_priv) >= 7)
> > -			dev_priv->display.load_luts = ivb_load_luts;
> > +			i915->display.color_commit = ilk_color_commit;
> > +
> > +		if (INTEL_GEN(i915) >= 11)
> > +			i915->display.load_luts = icl_load_luts;
> > +		else if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> > +			i915->display.load_luts = glk_load_luts;
> > +		else if (INTEL_GEN(i915) >= 8)
> > +			i915->display.load_luts = bdw_load_luts;
> > +		else if (INTEL_GEN(i915) >= 7)
> > +			i915->display.load_luts = ivb_load_luts;
> >   		else
> > -			dev_priv->display.load_luts = ilk_load_luts;
> > +			i915->display.load_luts = ilk_load_luts;
> >   	}
> >   	drm_crtc_enable_color_mgmt(&crtc->base,
> > -				   INTEL_INFO(dev_priv)->color.degamma_lut_size,
> > +				   INTEL_INFO(i915)->color.degamma_lut_size,
> >   				   has_ctm,
> > -				   INTEL_INFO(dev_priv)->color.gamma_lut_size);
> > +				   INTEL_INFO(i915)->color.gamma_lut_size);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_combo_phy.c b/drivers/gpu/drm/i915/intel_combo_phy.c
> > index 841708da5a56..0e7c18d783ca 100644
> > --- a/drivers/gpu/drm/i915/intel_combo_phy.c
> > +++ b/drivers/gpu/drm/i915/intel_combo_phy.c
> > @@ -44,7 +44,7 @@ static const struct cnl_procmon {
> >    * on its name.
> >    */
> >   static const struct cnl_procmon *
> > -cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port)
> > +cnl_get_procmon_ref_values(struct drm_i915_private *i915, enum port port)
> >   {
> >   	const struct cnl_procmon *procmon;
> >   	u32 val;
> > @@ -74,13 +74,13 @@ cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port)
> >   	return procmon;
> >   }
> > -static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv,
> > +static void cnl_set_procmon_ref_values(struct drm_i915_private *i915,
> >   				       enum port port)
> >   {
> >   	const struct cnl_procmon *procmon;
> >   	u32 val;
> > -	procmon = cnl_get_procmon_ref_values(dev_priv, port);
> > +	procmon = cnl_get_procmon_ref_values(i915, port);
> >   	val = I915_READ(ICL_PORT_COMP_DW1(port));
> >   	val &= ~((0xff << 16) | 0xff);
> > @@ -91,7 +91,7 @@ static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(ICL_PORT_COMP_DW10(port), procmon->dw10);
> >   }
> > -static bool check_phy_reg(struct drm_i915_private *dev_priv,
> > +static bool check_phy_reg(struct drm_i915_private *i915,
> >   			  enum port port, i915_reg_t reg, u32 mask,
> >   			  u32 expected_val)
> >   {
> > @@ -108,47 +108,47 @@ static bool check_phy_reg(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static bool cnl_verify_procmon_ref_values(struct drm_i915_private *dev_priv,
> > +static bool cnl_verify_procmon_ref_values(struct drm_i915_private *i915,
> >   					  enum port port)
> >   {
> >   	const struct cnl_procmon *procmon;
> >   	bool ret;
> > -	procmon = cnl_get_procmon_ref_values(dev_priv, port);
> > +	procmon = cnl_get_procmon_ref_values(i915, port);
> > -	ret = check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW1(port),
> > +	ret = check_phy_reg(i915, port, ICL_PORT_COMP_DW1(port),
> >   			    (0xff << 16) | 0xff, procmon->dw1);
> > -	ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW9(port),
> > +	ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW9(port),
> >   			     -1U, procmon->dw9);
> > -	ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW10(port),
> > +	ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW10(port),
> >   			     -1U, procmon->dw10);
> >   	return ret;
> >   }
> > -static bool cnl_combo_phy_enabled(struct drm_i915_private *dev_priv)
> > +static bool cnl_combo_phy_enabled(struct drm_i915_private *i915)
> >   {
> >   	return !(I915_READ(CHICKEN_MISC_2) & CNL_COMP_PWR_DOWN) &&
> >   		(I915_READ(CNL_PORT_COMP_DW0) & COMP_INIT);
> >   }
> > -static bool cnl_combo_phy_verify_state(struct drm_i915_private *dev_priv)
> > +static bool cnl_combo_phy_verify_state(struct drm_i915_private *i915)
> >   {
> >   	enum port port = PORT_A;
> >   	bool ret;
> > -	if (!cnl_combo_phy_enabled(dev_priv))
> > +	if (!cnl_combo_phy_enabled(i915))
> >   		return false;
> > -	ret = cnl_verify_procmon_ref_values(dev_priv, port);
> > +	ret = cnl_verify_procmon_ref_values(i915, port);
> > -	ret &= check_phy_reg(dev_priv, port, CNL_PORT_CL1CM_DW5,
> > +	ret &= check_phy_reg(i915, port, CNL_PORT_CL1CM_DW5,
> >   			     CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE);
> >   	return ret;
> >   }
> > -static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> > +static void cnl_combo_phys_init(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -157,7 +157,7 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(CHICKEN_MISC_2, val);
> >   	/* Dummy PORT_A to get the correct CNL register from the ICL macro */
> > -	cnl_set_procmon_ref_values(dev_priv, PORT_A);
> > +	cnl_set_procmon_ref_values(i915, PORT_A);
> >   	val = I915_READ(CNL_PORT_COMP_DW0);
> >   	val |= COMP_INIT;
> > @@ -168,11 +168,11 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(CNL_PORT_CL1CM_DW5, val);
> >   }
> > -static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> > +static void cnl_combo_phys_uninit(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	if (!cnl_combo_phy_verify_state(dev_priv))
> > +	if (!cnl_combo_phy_verify_state(i915))
> >   		DRM_WARN("Combo PHY HW state changed unexpectedly.\n");
> >   	val = I915_READ(CHICKEN_MISC_2);
> > @@ -180,7 +180,7 @@ static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(CHICKEN_MISC_2, val);
> >   }
> > -static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv,
> > +static bool icl_combo_phy_enabled(struct drm_i915_private *i915,
> >   				  enum port port)
> >   {
> >   	return !(I915_READ(ICL_PHY_MISC(port)) &
> > @@ -188,27 +188,27 @@ static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv,
> >   		(I915_READ(ICL_PORT_COMP_DW0(port)) & COMP_INIT);
> >   }
> > -static bool icl_combo_phy_verify_state(struct drm_i915_private *dev_priv,
> > +static bool icl_combo_phy_verify_state(struct drm_i915_private *i915,
> >   				       enum port port)
> >   {
> >   	bool ret;
> > -	if (!icl_combo_phy_enabled(dev_priv, port))
> > +	if (!icl_combo_phy_enabled(i915, port))
> >   		return false;
> > -	ret = cnl_verify_procmon_ref_values(dev_priv, port);
> > +	ret = cnl_verify_procmon_ref_values(i915, port);
> >   	if (port == PORT_A)
> > -		ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW8(port),
> > +		ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW8(port),
> >   				     IREFGEN, IREFGEN);
> > -	ret &= check_phy_reg(dev_priv, port, ICL_PORT_CL_DW5(port),
> > +	ret &= check_phy_reg(i915, port, ICL_PORT_CL_DW5(port),
> >   			     CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE);
> >   	return ret;
> >   }
> > -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> > +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915,
> >   				    enum port port, bool is_dsi,
> >   				    int lane_count, bool lane_reversal)
> >   {
> > @@ -260,14 +260,14 @@ void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(ICL_PORT_CL_DW10(port), val);
> >   }
> > -static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> > +static void icl_combo_phys_init(struct drm_i915_private *i915)
> >   {
> >   	enum port port;
> > -	for_each_combo_port(dev_priv, port) {
> > +	for_each_combo_port(i915, port) {
> >   		u32 val;
> > -		if (icl_combo_phy_verify_state(dev_priv, port)) {
> > +		if (icl_combo_phy_verify_state(i915, port)) {
> >   			DRM_DEBUG_DRIVER("Port %c combo PHY already enabled, won't reprogram it.\n",
> >   					 port_name(port));
> >   			continue;
> > @@ -277,7 +277,7 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> >   		val &= ~ICL_PHY_MISC_DE_IO_COMP_PWR_DOWN;
> >   		I915_WRITE(ICL_PHY_MISC(port), val);
> > -		cnl_set_procmon_ref_values(dev_priv, port);
> > +		cnl_set_procmon_ref_values(i915, port);
> >   		if (port == PORT_A) {
> >   			val = I915_READ(ICL_PORT_COMP_DW8(port));
> > @@ -295,15 +295,15 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void icl_combo_phys_uninit(struct drm_i915_private *dev_priv)
> > +static void icl_combo_phys_uninit(struct drm_i915_private *i915)
> >   {
> >   	enum port port;
> > -	for_each_combo_port_reverse(dev_priv, port) {
> > +	for_each_combo_port_reverse(i915, port) {
> >   		u32 val;
> >   		if (port == PORT_A &&
> > -		    !icl_combo_phy_verify_state(dev_priv, port))
> > +		    !icl_combo_phy_verify_state(i915, port))
> >   			DRM_WARN("Port %c combo PHY HW state changed unexpectedly\n",
> >   				 port_name(port));
> > diff --git a/drivers/gpu/drm/i915/intel_combo_phy.h b/drivers/gpu/drm/i915/intel_combo_phy.h
> > index e6e195a83b19..a08286cfcecf 100644
> > --- a/drivers/gpu/drm/i915/intel_combo_phy.h
> > +++ b/drivers/gpu/drm/i915/intel_combo_phy.h
> > @@ -11,9 +11,9 @@
> >   struct drm_i915_private;
> > -void intel_combo_phy_init(struct drm_i915_private *dev_priv);
> > -void intel_combo_phy_uninit(struct drm_i915_private *dev_priv);
> > -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv,
> > +void intel_combo_phy_init(struct drm_i915_private *i915);
> > +void intel_combo_phy_uninit(struct drm_i915_private *i915);
> > +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915,
> >   				    enum port port, bool is_dsi,
> >   				    int lane_count, bool lane_reversal);
> > diff --git a/drivers/gpu/drm/i915/intel_connector.c b/drivers/gpu/drm/i915/intel_connector.c
> > index 073b6c3ab7cc..2912804d5a0d 100644
> > --- a/drivers/gpu/drm/i915/intel_connector.c
> > +++ b/drivers/gpu/drm/i915/intel_connector.c
> > @@ -219,10 +219,10 @@ void
> >   intel_attach_force_audio_property(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_property *prop;
> > -	prop = dev_priv->force_audio_property;
> > +	prop = i915->force_audio_property;
> >   	if (prop == NULL) {
> >   		prop = drm_property_create_enum(dev, 0,
> >   					   "audio",
> > @@ -231,7 +231,7 @@ intel_attach_force_audio_property(struct drm_connector *connector)
> >   		if (prop == NULL)
> >   			return;
> > -		dev_priv->force_audio_property = prop;
> > +		i915->force_audio_property = prop;
> >   	}
> >   	drm_object_attach_property(&connector->base, prop, 0);
> >   }
> > @@ -246,10 +246,10 @@ void
> >   intel_attach_broadcast_rgb_property(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_property *prop;
> > -	prop = dev_priv->broadcast_rgb_property;
> > +	prop = i915->broadcast_rgb_property;
> >   	if (prop == NULL) {
> >   		prop = drm_property_create_enum(dev, DRM_MODE_PROP_ENUM,
> >   					   "Broadcast RGB",
> > @@ -258,7 +258,7 @@ intel_attach_broadcast_rgb_property(struct drm_connector *connector)
> >   		if (prop == NULL)
> >   			return;
> > -		dev_priv->broadcast_rgb_property = prop;
> > +		i915->broadcast_rgb_property = prop;
> >   	}
> >   	drm_object_attach_property(&connector->base, prop, 0);
> > diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c
> > index 3fcf2f84bcce..2a794e0bf457 100644
> > --- a/drivers/gpu/drm/i915/intel_crt.c
> > +++ b/drivers/gpu/drm/i915/intel_crt.c
> > @@ -70,7 +70,7 @@ static struct intel_crt *intel_attached_crt(struct drm_connector *connector)
> >   	return intel_encoder_to_crt(intel_attached_encoder(connector));
> >   }
> > -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_crt_port_enabled(struct drm_i915_private *i915,
> >   			    i915_reg_t adpa_reg, enum pipe *pipe)
> >   {
> >   	u32 val;
> > @@ -78,7 +78,7 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(adpa_reg);
> >   	/* asserts want to know the pipe even if the port is disabled */
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		*pipe = (val & ADPA_PIPE_SEL_MASK_CPT) >> ADPA_PIPE_SEL_SHIFT_CPT;
> >   	else
> >   		*pipe = (val & ADPA_PIPE_SEL_MASK) >> ADPA_PIPE_SEL_SHIFT;
> > @@ -89,26 +89,26 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> >   static bool intel_crt_get_hw_state(struct intel_encoder *encoder,
> >   				   enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crt *crt = intel_encoder_to_crt(encoder);
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > -	ret = intel_crt_port_enabled(dev_priv, crt->adpa_reg, pipe);
> > +	ret = intel_crt_port_enabled(i915, crt->adpa_reg, pipe);
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> >   static unsigned int intel_crt_get_flags(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crt *crt = intel_encoder_to_crt(encoder);
> >   	u32 tmp, flags = 0;
> > @@ -140,7 +140,7 @@ static void intel_crt_get_config(struct intel_encoder *encoder,
> >   static void hsw_crt_get_config(struct intel_encoder *encoder,
> >   			       struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	intel_ddi_get_config(encoder, pipe_config);
> > @@ -150,7 +150,7 @@ static void hsw_crt_get_config(struct intel_encoder *encoder,
> >   					      DRM_MODE_FLAG_NVSYNC);
> >   	pipe_config->base.adjusted_mode.flags |= intel_crt_get_flags(encoder);
> > -	pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(dev_priv);
> > +	pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(i915);
> >   }
> >   /* Note: The caller is required to filter out dpms modes not supported by the
> > @@ -159,13 +159,13 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       int mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crt *crt = intel_encoder_to_crt(encoder);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> >   	u32 adpa;
> > -	if (INTEL_GEN(dev_priv) >= 5)
> > +	if (INTEL_GEN(i915) >= 5)
> >   		adpa = ADPA_HOTPLUG_BITS;
> >   	else
> >   		adpa = 0;
> > @@ -176,14 +176,14 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder,
> >   		adpa |= ADPA_VSYNC_ACTIVE_HIGH;
> >   	/* For CPT allow 3 pipe config, for others just use A or B */
> > -	if (HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_LPT(i915))
> >   		; /* Those bits don't exist here */
> > -	else if (HAS_PCH_CPT(dev_priv))
> > +	else if (HAS_PCH_CPT(i915))
> >   		adpa |= ADPA_PIPE_SEL_CPT(crtc->pipe);
> >   	else
> >   		adpa |= ADPA_PIPE_SEL(crtc->pipe);
> > -	if (!HAS_PCH_SPLIT(dev_priv))
> > +	if (!HAS_PCH_SPLIT(i915))
> >   		I915_WRITE(BCLRPAT(crtc->pipe), 0);
> >   	switch (mode) {
> > @@ -228,57 +228,57 @@ static void hsw_disable_crt(struct intel_encoder *encoder,
> >   			    const struct intel_crtc_state *old_crtc_state,
> >   			    const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	WARN_ON(!old_crtc_state->has_pch_encoder);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > +	intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> >   }
> >   static void hsw_post_disable_crt(struct intel_encoder *encoder,
> >   				 const struct intel_crtc_state *old_crtc_state,
> >   				 const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	intel_ddi_disable_pipe_clock(old_crtc_state);
> >   	pch_post_disable_crt(encoder, old_crtc_state, old_conn_state);
> > -	lpt_disable_pch_transcoder(dev_priv);
> > -	lpt_disable_iclkip(dev_priv);
> > +	lpt_disable_pch_transcoder(i915);
> > +	lpt_disable_iclkip(i915);
> >   	intel_ddi_fdi_post_disable(encoder, old_crtc_state, old_conn_state);
> >   	WARN_ON(!old_crtc_state->has_pch_encoder);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > +	intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> >   }
> >   static void hsw_pre_pll_enable_crt(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *crtc_state,
> >   				   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	WARN_ON(!crtc_state->has_pch_encoder);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > +	intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> >   }
> >   static void hsw_pre_enable_crt(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum pipe pipe = crtc->pipe;
> >   	WARN_ON(!crtc_state->has_pch_encoder);
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> > -	dev_priv->display.fdi_link_train(crtc, crtc_state);
> > +	i915->display.fdi_link_train(crtc, crtc_state);
> >   	intel_ddi_enable_pipe_clock(crtc_state);
> >   }
> > @@ -287,7 +287,7 @@ static void hsw_enable_crt(struct intel_encoder *encoder,
> >   			   const struct intel_crtc_state *crtc_state,
> >   			   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -295,10 +295,10 @@ static void hsw_enable_crt(struct intel_encoder *encoder,
> >   	intel_crt_set_dpms(encoder, crtc_state, DRM_MODE_DPMS_ON);
> > -	intel_wait_for_vblank(dev_priv, pipe);
> > -	intel_wait_for_vblank(dev_priv, pipe);
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > +	intel_wait_for_vblank(i915, pipe);
> > +	intel_wait_for_vblank(i915, pipe);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> > +	intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> >   }
> >   static void intel_enable_crt(struct intel_encoder *encoder,
> > @@ -313,8 +313,8 @@ intel_crt_mode_valid(struct drm_connector *connector,
> >   		     struct drm_display_mode *mode)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	int max_dotclk = dev_priv->max_dotclk_freq;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	int max_dotclk = i915->max_dotclk_freq;
> >   	int max_clock;
> >   	if (mode->flags & DRM_MODE_FLAG_DBLSCAN)
> > @@ -323,15 +323,15 @@ intel_crt_mode_valid(struct drm_connector *connector,
> >   	if (mode->clock < 25000)
> >   		return MODE_CLOCK_LOW;
> > -	if (HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_LPT(i915))
> >   		max_clock = 180000;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		/*
> >   		 * 270 MHz due to current DPLL limits,
> >   		 * DAC limit supposedly 355 MHz.
> >   		 */
> >   		max_clock = 270000;
> > -	else if (IS_GEN_RANGE(dev_priv, 3, 4))
> > +	else if (IS_GEN_RANGE(i915, 3, 4))
> >   		max_clock = 400000;
> >   	else
> >   		max_clock = 350000;
> > @@ -342,7 +342,7 @@ intel_crt_mode_valid(struct drm_connector *connector,
> >   		return MODE_CLOCK_HIGH;
> >   	/* The FDI receiver on LPT only supports 8bpc and only has 2 lanes. */
> > -	if (HAS_PCH_LPT(dev_priv) &&
> > +	if (HAS_PCH_LPT(i915) &&
> >   	    (ironlake_get_lanes_required(mode->clock, 270000, 24) > 2))
> >   		return MODE_CLOCK_HIGH;
> > @@ -388,7 +388,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder,
> >   				  struct intel_crtc_state *pipe_config,
> >   				  struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_display_mode *adjusted_mode =
> >   		&pipe_config->base.adjusted_mode;
> > @@ -404,7 +404,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder,
> >   	pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB;
> >   	/* LPT FDI RX only supports 8bpc. */
> > -	if (HAS_PCH_LPT(dev_priv)) {
> > +	if (HAS_PCH_LPT(i915)) {
> >   		if (pipe_config->bw_constrained && pipe_config->pipe_bpp < 24) {
> >   			DRM_DEBUG_KMS("LPT only supports 24bpp\n");
> >   			return -EINVAL;
> > @@ -423,13 +423,13 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> >   	struct intel_crt *crt = intel_attached_crt(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 adpa;
> >   	bool ret;
> >   	/* The first time through, trigger an explicit detection cycle */
> >   	if (crt->force_hotplug_required) {
> > -		bool turn_off_dac = HAS_PCH_SPLIT(dev_priv);
> > +		bool turn_off_dac = HAS_PCH_SPLIT(i915);
> >   		u32 save_adpa;
> >   		crt->force_hotplug_required = 0;
> > @@ -443,7 +443,7 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector)
> >   		I915_WRITE(crt->adpa_reg, adpa);
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    crt->adpa_reg,
> >   					    ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0,
> >   					    1000))
> > @@ -470,7 +470,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> >   	struct intel_crt *crt = intel_attached_crt(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	bool reenable_hpd;
> >   	u32 adpa;
> >   	bool ret;
> > @@ -488,7 +488,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> >   	 *
> >   	 * Just disable HPD interrupts here to prevent this
> >   	 */
> > -	reenable_hpd = intel_hpd_disable(dev_priv, crt->base.hpd_pin);
> > +	reenable_hpd = intel_hpd_disable(i915, crt->base.hpd_pin);
> >   	save_adpa = adpa = I915_READ(crt->adpa_reg);
> >   	DRM_DEBUG_KMS("trigger hotplug detect cycle: adpa=0x%x\n", adpa);
> > @@ -497,7 +497,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> >   	I915_WRITE(crt->adpa_reg, adpa);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    crt->adpa_reg,
> >   				    ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0,
> >   				    1000)) {
> > @@ -515,7 +515,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> >   	DRM_DEBUG_KMS("valleyview hotplug adpa=0x%x, result %d\n", adpa, ret);
> >   	if (reenable_hpd)
> > -		intel_hpd_enable(dev_priv, crt->base.hpd_pin);
> > +		intel_hpd_enable(i915, crt->base.hpd_pin);
> >   	return ret;
> >   }
> > @@ -523,15 +523,15 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector)
> >   static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 stat;
> >   	bool ret = false;
> >   	int i, tries = 0;
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		return intel_ironlake_crt_detect_hotplug(connector);
> > -	if (IS_VALLEYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915))
> >   		return valleyview_crt_detect_hotplug(connector);
> >   	/*
> > @@ -539,18 +539,18 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> >   	 * to get a reliable result.
> >   	 */
> > -	if (IS_G45(dev_priv))
> > +	if (IS_G45(i915))
> >   		tries = 2;
> >   	else
> >   		tries = 1;
> >   	for (i = 0; i < tries ; i++) {
> >   		/* turn on the FORCE_DETECT */
> > -		i915_hotplug_interrupt_update(dev_priv,
> > +		i915_hotplug_interrupt_update(i915,
> >   					      CRT_HOTPLUG_FORCE_DETECT,
> >   					      CRT_HOTPLUG_FORCE_DETECT);
> >   		/* wait for FORCE_DETECT to go off */
> > -		if (intel_wait_for_register(&dev_priv->uncore, PORT_HOTPLUG_EN,
> > +		if (intel_wait_for_register(&i915->uncore, PORT_HOTPLUG_EN,
> >   					    CRT_HOTPLUG_FORCE_DETECT, 0,
> >   					    1000))
> >   			DRM_DEBUG_KMS("timed out waiting for FORCE_DETECT to go off");
> > @@ -563,7 +563,7 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector)
> >   	/* clear the interrupt we just generated, if any */
> >   	I915_WRITE(PORT_HOTPLUG_STAT, CRT_HOTPLUG_INT_STATUS);
> > -	i915_hotplug_interrupt_update(dev_priv, CRT_HOTPLUG_FORCE_DETECT, 0);
> > +	i915_hotplug_interrupt_update(i915, CRT_HOTPLUG_FORCE_DETECT, 0);
> >   	return ret;
> >   }
> > @@ -605,14 +605,14 @@ static int intel_crt_ddc_get_modes(struct drm_connector *connector,
> >   static bool intel_crt_detect_ddc(struct drm_connector *connector)
> >   {
> >   	struct intel_crt *crt = intel_attached_crt(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(crt->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crt->base.base.dev);
> >   	struct edid *edid;
> >   	struct i2c_adapter *i2c;
> >   	bool ret = false;
> >   	BUG_ON(crt->base.type != INTEL_OUTPUT_ANALOG);
> > -	i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin);
> > +	i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin);
> >   	edid = intel_crt_get_edid(connector, i2c);
> >   	if (edid) {
> > @@ -642,8 +642,8 @@ static enum drm_connector_status
> >   intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> >   {
> >   	struct drm_device *dev = crt->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 save_bclrpat;
> >   	u32 save_vtotal;
> >   	u32 vtotal, vactive;
> > @@ -677,7 +677,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> >   	/* Set the border color to purple. */
> >   	intel_uncore_write(uncore, bclrpat_reg, 0x500050);
> > -	if (!IS_GEN(dev_priv, 2)) {
> > +	if (!IS_GEN(i915, 2)) {
> >   		u32 pipeconf = intel_uncore_read(uncore, pipeconf_reg);
> >   		intel_uncore_write(uncore,
> >   				   pipeconf_reg,
> > @@ -685,7 +685,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe)
> >   		intel_uncore_posting_read(uncore, pipeconf_reg);
> >   		/* Wait for next Vblank to substitue
> >   		 * border color for Color info */
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> >   		st00 = intel_uncore_read8(uncore, _VGA_MSR_WRITE);
> >   		status = ((st00 & (1 << 4)) != 0) ?
> >   			connector_status_connected :
> > @@ -789,7 +789,7 @@ intel_crt_detect(struct drm_connector *connector,
> >   		 struct drm_modeset_acquire_ctx *ctx,
> >   		 bool force)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_crt *crt = intel_attached_crt(connector);
> >   	struct intel_encoder *intel_encoder = &crt->base;
> >   	intel_wakeref_t wakeref;
> > @@ -801,7 +801,7 @@ intel_crt_detect(struct drm_connector *connector,
> >   		      force);
> >   	if (i915_modparams.load_detect_test) {
> > -		wakeref = intel_display_power_get(dev_priv,
> > +		wakeref = intel_display_power_get(i915,
> >   						  intel_encoder->power_domain);
> >   		goto load_detect;
> >   	}
> > @@ -810,10 +810,10 @@ intel_crt_detect(struct drm_connector *connector,
> >   	if (dmi_check_system(intel_spurious_crt_detect))
> >   		return connector_status_disconnected;
> > -	wakeref = intel_display_power_get(dev_priv,
> > +	wakeref = intel_display_power_get(i915,
> >   					  intel_encoder->power_domain);
> > -	if (I915_HAS_HOTPLUG(dev_priv)) {
> > +	if (I915_HAS_HOTPLUG(i915)) {
> >   		/* We can not rely on the HPD pin always being correctly wired
> >   		 * up, for example many KVM do not pass it through, and so
> >   		 * only trust an assertion that the monitor is connected.
> > @@ -835,7 +835,7 @@ intel_crt_detect(struct drm_connector *connector,
> >   	 * broken monitor (without edid) to work behind a broken kvm (that fails
> >   	 * to have the right resistors for HP detection) needs to fix this up.
> >   	 * For now just bail out. */
> > -	if (I915_HAS_HOTPLUG(dev_priv)) {
> > +	if (I915_HAS_HOTPLUG(i915)) {
> >   		status = connector_status_disconnected;
> >   		goto out;
> >   	}
> > @@ -851,7 +851,7 @@ intel_crt_detect(struct drm_connector *connector,
> >   	if (ret > 0) {
> >   		if (intel_crt_detect_ddc(connector))
> >   			status = connector_status_connected;
> > -		else if (INTEL_GEN(dev_priv) < 4)
> > +		else if (INTEL_GEN(i915) < 4)
> >   			status = intel_crt_load_detect(crt,
> >   				to_intel_crtc(connector->state->crtc)->pipe);
> >   		else if (i915_modparams.load_detect_test)
> > @@ -866,44 +866,44 @@ intel_crt_detect(struct drm_connector *connector,
> >   	}
> >   out:
> > -	intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
> >   	return status;
> >   }
> >   static int intel_crt_get_modes(struct drm_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crt *crt = intel_attached_crt(connector);
> >   	struct intel_encoder *intel_encoder = &crt->base;
> >   	intel_wakeref_t wakeref;
> >   	struct i2c_adapter *i2c;
> >   	int ret;
> > -	wakeref = intel_display_power_get(dev_priv,
> > +	wakeref = intel_display_power_get(i915,
> >   					  intel_encoder->power_domain);
> > -	i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin);
> > +	i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin);
> >   	ret = intel_crt_ddc_get_modes(connector, i2c);
> > -	if (ret || !IS_G4X(dev_priv))
> > +	if (ret || !IS_G4X(i915))
> >   		goto out;
> >   	/* Try to probe digital port for output in DVI-I -> VGA mode. */
> > -	i2c = intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPB);
> > +	i2c = intel_gmbus_get_adapter(i915, GMBUS_PIN_DPB);
> >   	ret = intel_crt_ddc_get_modes(connector, i2c);
> >   out:
> > -	intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> >   void intel_crt_reset(struct drm_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->dev);
> >   	struct intel_crt *crt = intel_encoder_to_crt(to_intel_encoder(encoder));
> > -	if (INTEL_GEN(dev_priv) >= 5) {
> > +	if (INTEL_GEN(i915) >= 5) {
> >   		u32 adpa;
> >   		adpa = I915_READ(crt->adpa_reg);
> > @@ -942,7 +942,7 @@ static const struct drm_encoder_funcs intel_crt_enc_funcs = {
> >   	.destroy = intel_encoder_destroy,
> >   };
> > -void intel_crt_init(struct drm_i915_private *dev_priv)
> > +void intel_crt_init(struct drm_i915_private *i915)
> >   {
> >   	struct drm_connector *connector;
> >   	struct intel_crt *crt;
> > @@ -950,9 +950,9 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   	i915_reg_t adpa_reg;
> >   	u32 adpa;
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		adpa_reg = PCH_ADPA;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		adpa_reg = VLV_ADPA;
> >   	else
> >   		adpa_reg = ADPA;
> > @@ -986,22 +986,22 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   	connector = &intel_connector->base;
> >   	crt->connector = intel_connector;
> > -	drm_connector_init(&dev_priv->drm, &intel_connector->base,
> > +	drm_connector_init(&i915->drm, &intel_connector->base,
> >   			   &intel_crt_connector_funcs, DRM_MODE_CONNECTOR_VGA);
> > -	drm_encoder_init(&dev_priv->drm, &crt->base.base, &intel_crt_enc_funcs,
> > +	drm_encoder_init(&i915->drm, &crt->base.base, &intel_crt_enc_funcs,
> >   			 DRM_MODE_ENCODER_DAC, "CRT");
> >   	intel_connector_attach_encoder(intel_connector, &crt->base);
> >   	crt->base.type = INTEL_OUTPUT_ANALOG;
> >   	crt->base.cloneable = (1 << INTEL_OUTPUT_DVO) | (1 << INTEL_OUTPUT_HDMI);
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		crt->base.crtc_mask = (1 << 0);
> >   	else
> >   		crt->base.crtc_mask = (1 << 0) | (1 << 1) | (1 << 2);
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		connector->interlace_allowed = 0;
> >   	else
> >   		connector->interlace_allowed = 1;
> > @@ -1011,13 +1011,13 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   	crt->base.power_domain = POWER_DOMAIN_PORT_CRT;
> > -	if (I915_HAS_HOTPLUG(dev_priv) &&
> > +	if (I915_HAS_HOTPLUG(i915) &&
> >   	    !dmi_check_system(intel_spurious_crt_detect)) {
> >   		crt->base.hpd_pin = HPD_CRT;
> >   		crt->base.hotplug = intel_encoder_hotplug;
> >   	}
> > -	if (HAS_DDI(dev_priv)) {
> > +	if (HAS_DDI(i915)) {
> >   		crt->base.port = PORT_E;
> >   		crt->base.get_config = hsw_crt_get_config;
> >   		crt->base.get_hw_state = intel_ddi_get_hw_state;
> > @@ -1028,7 +1028,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   		crt->base.disable = hsw_disable_crt;
> >   		crt->base.post_disable = hsw_post_disable_crt;
> >   	} else {
> > -		if (HAS_PCH_SPLIT(dev_priv)) {
> > +		if (HAS_PCH_SPLIT(i915)) {
> >   			crt->base.compute_config = pch_crt_compute_config;
> >   			crt->base.disable = pch_disable_crt;
> >   			crt->base.post_disable = pch_post_disable_crt;
> > @@ -1045,7 +1045,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   	drm_connector_helper_add(connector, &intel_crt_connector_helper_funcs);
> > -	if (!I915_HAS_HOTPLUG(dev_priv))
> > +	if (!I915_HAS_HOTPLUG(i915))
> >   		intel_connector->polled = DRM_CONNECTOR_POLL_CONNECT;
> >   	/*
> > @@ -1058,11 +1058,11 @@ void intel_crt_init(struct drm_i915_private *dev_priv)
> >   	 * polarity and link reversal bits or not, instead of relying on the
> >   	 * BIOS.
> >   	 */
> > -	if (HAS_PCH_LPT(dev_priv)) {
> > +	if (HAS_PCH_LPT(i915)) {
> >   		u32 fdi_config = FDI_RX_POLARITY_REVERSED_LPT |
> >   				 FDI_RX_LINK_REVERSAL_OVERRIDE;
> > -		dev_priv->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config;
> > +		i915->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config;
> >   	}
> >   	intel_crt_reset(&crt->base.base);
> > diff --git a/drivers/gpu/drm/i915/intel_crt.h b/drivers/gpu/drm/i915/intel_crt.h
> > index 1b3fba359efc..f771fbcf8b27 100644
> > --- a/drivers/gpu/drm/i915/intel_crt.h
> > +++ b/drivers/gpu/drm/i915/intel_crt.h
> > @@ -13,9 +13,9 @@ struct drm_encoder;
> >   struct drm_i915_private;
> >   struct drm_i915_private;
> > -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_crt_port_enabled(struct drm_i915_private *i915,
> >   			    i915_reg_t adpa_reg, enum pipe *pipe);
> > -void intel_crt_init(struct drm_i915_private *dev_priv);
> > +void intel_crt_init(struct drm_i915_private *i915);
> >   void intel_crt_reset(struct drm_encoder *encoder);
> >   #endif /* __INTEL_CRT_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_csr.c b/drivers/gpu/drm/i915/intel_csr.c
> > index bf0eebd385b9..29ea3cb6bbaf 100644
> > --- a/drivers/gpu/drm/i915/intel_csr.c
> > +++ b/drivers/gpu/drm/i915/intel_csr.c
> > @@ -206,18 +206,18 @@ static const struct stepping_info icl_stepping_info[] = {
> >   static const struct stepping_info no_stepping_info = { '*', '*' };
> >   static const struct stepping_info *
> > -intel_get_stepping_info(struct drm_i915_private *dev_priv)
> > +intel_get_stepping_info(struct drm_i915_private *i915)
> >   {
> >   	const struct stepping_info *si;
> >   	unsigned int size;
> > -	if (IS_ICELAKE(dev_priv)) {
> > +	if (IS_ICELAKE(i915)) {
> >   		size = ARRAY_SIZE(icl_stepping_info);
> >   		si = icl_stepping_info;
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)) {
> >   		size = ARRAY_SIZE(skl_stepping_info);
> >   		si = skl_stepping_info;
> > -	} else if (IS_BROXTON(dev_priv)) {
> > +	} else if (IS_BROXTON(i915)) {
> >   		size = ARRAY_SIZE(bxt_stepping_info);
> >   		si = bxt_stepping_info;
> >   	} else {
> > @@ -225,19 +225,19 @@ intel_get_stepping_info(struct drm_i915_private *dev_priv)
> >   		si = NULL;
> >   	}
> > -	if (INTEL_REVID(dev_priv) < size)
> > -		return si + INTEL_REVID(dev_priv);
> > +	if (INTEL_REVID(i915) < size)
> > +		return si + INTEL_REVID(i915);
> >   	return &no_stepping_info;
> >   }
> > -static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv)
> > +static void gen9_set_dc_state_debugmask(struct drm_i915_private *i915)
> >   {
> >   	u32 val, mask;
> >   	mask = DC_STATE_DEBUG_MASK_MEMORY_UP;
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		mask |= DC_STATE_DEBUG_MASK_CORES;
> >   	/* The below bit doesn't need to be cleared ever afterwards */
> > @@ -251,29 +251,29 @@ static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_csr_load_program() - write the firmware from memory to register.
> > - * @dev_priv: i915 drm device.
> > + * @i915: i915 drm device.
> >    *
> >    * CSR firmware is read from a .bin file and kept in internal memory one time.
> >    * Everytime display comes back from low power state this function is called to
> >    * copy the firmware from internal memory to registers.
> >    */
> > -void intel_csr_load_program(struct drm_i915_private *dev_priv)
> > +void intel_csr_load_program(struct drm_i915_private *i915)
> >   {
> > -	u32 *payload = dev_priv->csr.dmc_payload;
> > +	u32 *payload = i915->csr.dmc_payload;
> >   	u32 i, fw_size;
> > -	if (!HAS_CSR(dev_priv)) {
> > +	if (!HAS_CSR(i915)) {
> >   		DRM_ERROR("No CSR support available for this platform\n");
> >   		return;
> >   	}
> > -	if (!dev_priv->csr.dmc_payload) {
> > +	if (!i915->csr.dmc_payload) {
> >   		DRM_ERROR("Tried to program CSR with empty payload\n");
> >   		return;
> >   	}
> > -	fw_size = dev_priv->csr.dmc_fw_size;
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	fw_size = i915->csr.dmc_fw_size;
> > +	assert_rpm_wakelock_held(i915);
> >   	preempt_disable();
> > @@ -282,24 +282,24 @@ void intel_csr_load_program(struct drm_i915_private *dev_priv)
> >   	preempt_enable();
> > -	for (i = 0; i < dev_priv->csr.mmio_count; i++) {
> > -		I915_WRITE(dev_priv->csr.mmioaddr[i],
> > -			   dev_priv->csr.mmiodata[i]);
> > +	for (i = 0; i < i915->csr.mmio_count; i++) {
> > +		I915_WRITE(i915->csr.mmioaddr[i],
> > +			   i915->csr.mmiodata[i]);
> >   	}
> > -	dev_priv->csr.dc_state = 0;
> > +	i915->csr.dc_state = 0;
> > -	gen9_set_dc_state_debugmask(dev_priv);
> > +	gen9_set_dc_state_debugmask(i915);
> >   }
> > -static u32 *parse_csr_fw(struct drm_i915_private *dev_priv,
> > +static u32 *parse_csr_fw(struct drm_i915_private *i915,
> >   			 const struct firmware *fw)
> >   {
> >   	struct intel_css_header *css_header;
> >   	struct intel_package_header *package_header;
> >   	struct intel_dmc_header *dmc_header;
> > -	struct intel_csr *csr = &dev_priv->csr;
> > -	const struct stepping_info *si = intel_get_stepping_info(dev_priv);
> > +	struct intel_csr *csr = &i915->csr;
> > +	const struct stepping_info *si = intel_get_stepping_info(i915);
> >   	u32 dmc_offset = CSR_DEFAULT_FW_OFFSET, readcount = 0, nbytes;
> >   	u32 i;
> >   	u32 *dmc_payload;
> > @@ -430,48 +430,48 @@ static u32 *parse_csr_fw(struct drm_i915_private *dev_priv,
> >   	return NULL;
> >   }
> > -static void intel_csr_runtime_pm_get(struct drm_i915_private *dev_priv)
> > +static void intel_csr_runtime_pm_get(struct drm_i915_private *i915)
> >   {
> > -	WARN_ON(dev_priv->csr.wakeref);
> > -	dev_priv->csr.wakeref =
> > -		intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> > +	WARN_ON(i915->csr.wakeref);
> > +	i915->csr.wakeref =
> > +		intel_display_power_get(i915, POWER_DOMAIN_INIT);
> >   }
> > -static void intel_csr_runtime_pm_put(struct drm_i915_private *dev_priv)
> > +static void intel_csr_runtime_pm_put(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref __maybe_unused =
> > -		fetch_and_zero(&dev_priv->csr.wakeref);
> > +		fetch_and_zero(&i915->csr.wakeref);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
> >   }
> >   static void csr_load_work_fn(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	struct intel_csr *csr;
> >   	const struct firmware *fw = NULL;
> > -	dev_priv = container_of(work, typeof(*dev_priv), csr.work);
> > -	csr = &dev_priv->csr;
> > +	i915 = container_of(work, typeof(*i915), csr.work);
> > +	csr = &i915->csr;
> > -	request_firmware(&fw, dev_priv->csr.fw_path, &dev_priv->drm.pdev->dev);
> > +	request_firmware(&fw, i915->csr.fw_path, &i915->drm.pdev->dev);
> >   	if (fw)
> > -		dev_priv->csr.dmc_payload = parse_csr_fw(dev_priv, fw);
> > +		i915->csr.dmc_payload = parse_csr_fw(i915, fw);
> > -	if (dev_priv->csr.dmc_payload) {
> > -		intel_csr_load_program(dev_priv);
> > -		intel_csr_runtime_pm_put(dev_priv);
> > +	if (i915->csr.dmc_payload) {
> > +		intel_csr_load_program(i915);
> > +		intel_csr_runtime_pm_put(i915);
> >   		DRM_INFO("Finished loading DMC firmware %s (v%u.%u)\n",
> > -			 dev_priv->csr.fw_path,
> > +			 i915->csr.fw_path,
> >   			 CSR_VERSION_MAJOR(csr->version),
> >   			 CSR_VERSION_MINOR(csr->version));
> >   	} else {
> > -		dev_notice(dev_priv->drm.dev,
> > +		dev_notice(i915->drm.dev,
> >   			   "Failed to load DMC firmware %s."
> >   			   " Disabling runtime power management.\n",
> >   			   csr->fw_path);
> > -		dev_notice(dev_priv->drm.dev, "DMC firmware homepage: %s",
> > +		dev_notice(i915->drm.dev, "DMC firmware homepage: %s",
> >   			   INTEL_UC_FIRMWARE_URL);
> >   	}
> > @@ -480,18 +480,18 @@ static void csr_load_work_fn(struct work_struct *work)
> >   /**
> >    * intel_csr_ucode_init() - initialize the firmware loading.
> > - * @dev_priv: i915 drm device.
> > + * @i915: i915 drm device.
> >    *
> >    * This function is called at the time of loading the display driver to read
> >    * firmware from a .bin file and copied into a internal memory.
> >    */
> > -void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> > +void intel_csr_ucode_init(struct drm_i915_private *i915)
> >   {
> > -	struct intel_csr *csr = &dev_priv->csr;
> > +	struct intel_csr *csr = &i915->csr;
> > -	INIT_WORK(&dev_priv->csr.work, csr_load_work_fn);
> > +	INIT_WORK(&i915->csr.work, csr_load_work_fn);
> > -	if (!HAS_CSR(dev_priv))
> > +	if (!HAS_CSR(i915))
> >   		return;
> >   	/*
> > @@ -502,32 +502,32 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> >   	 * suspend as runtime suspend *requires* a working CSR for whatever
> >   	 * reason.
> >   	 */
> > -	intel_csr_runtime_pm_get(dev_priv);
> > +	intel_csr_runtime_pm_get(i915);
> > -	if (INTEL_GEN(dev_priv) >= 12) {
> > +	if (INTEL_GEN(i915) >= 12) {
> >   		/* Allow to load fw via parameter using the last known size */
> >   		csr->max_fw_size = GEN12_CSR_MAX_FW_SIZE;
> > -	} else if (IS_GEN(dev_priv, 11)) {
> > +	} else if (IS_GEN(i915, 11)) {
> >   		csr->fw_path = ICL_CSR_PATH;
> >   		csr->required_version = ICL_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = ICL_CSR_MAX_FW_SIZE;
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_CANNONLAKE(i915)) {
> >   		csr->fw_path = CNL_CSR_PATH;
> >   		csr->required_version = CNL_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = CNL_CSR_MAX_FW_SIZE;
> > -	} else if (IS_GEMINILAKE(dev_priv)) {
> > +	} else if (IS_GEMINILAKE(i915)) {
> >   		csr->fw_path = GLK_CSR_PATH;
> >   		csr->required_version = GLK_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = GLK_CSR_MAX_FW_SIZE;
> > -	} else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> > +	} else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> >   		csr->fw_path = KBL_CSR_PATH;
> >   		csr->required_version = KBL_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = KBL_CSR_MAX_FW_SIZE;
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)) {
> >   		csr->fw_path = SKL_CSR_PATH;
> >   		csr->required_version = SKL_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = SKL_CSR_MAX_FW_SIZE;
> > -	} else if (IS_BROXTON(dev_priv)) {
> > +	} else if (IS_BROXTON(i915)) {
> >   		csr->fw_path = BXT_CSR_PATH;
> >   		csr->required_version = BXT_CSR_VERSION_REQUIRED;
> >   		csr->max_fw_size = BXT_CSR_MAX_FW_SIZE;
> > @@ -551,63 +551,63 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv)
> >   	}
> >   	DRM_DEBUG_KMS("Loading %s\n", csr->fw_path);
> > -	schedule_work(&dev_priv->csr.work);
> > +	schedule_work(&i915->csr.work);
> >   }
> >   /**
> >    * intel_csr_ucode_suspend() - prepare CSR firmware before system suspend
> > - * @dev_priv: i915 drm device
> > + * @i915: i915 drm device
> >    *
> >    * Prepare the DMC firmware before entering system suspend. This includes
> >    * flushing pending work items and releasing any resources acquired during
> >    * init.
> >    */
> > -void intel_csr_ucode_suspend(struct drm_i915_private *dev_priv)
> > +void intel_csr_ucode_suspend(struct drm_i915_private *i915)
> >   {
> > -	if (!HAS_CSR(dev_priv))
> > +	if (!HAS_CSR(i915))
> >   		return;
> > -	flush_work(&dev_priv->csr.work);
> > +	flush_work(&i915->csr.work);
> >   	/* Drop the reference held in case DMC isn't loaded. */
> > -	if (!dev_priv->csr.dmc_payload)
> > -		intel_csr_runtime_pm_put(dev_priv);
> > +	if (!i915->csr.dmc_payload)
> > +		intel_csr_runtime_pm_put(i915);
> >   }
> >   /**
> >    * intel_csr_ucode_resume() - init CSR firmware during system resume
> > - * @dev_priv: i915 drm device
> > + * @i915: i915 drm device
> >    *
> >    * Reinitialize the DMC firmware during system resume, reacquiring any
> >    * resources released in intel_csr_ucode_suspend().
> >    */
> > -void intel_csr_ucode_resume(struct drm_i915_private *dev_priv)
> > +void intel_csr_ucode_resume(struct drm_i915_private *i915)
> >   {
> > -	if (!HAS_CSR(dev_priv))
> > +	if (!HAS_CSR(i915))
> >   		return;
> >   	/*
> >   	 * Reacquire the reference to keep RPM disabled in case DMC isn't
> >   	 * loaded.
> >   	 */
> > -	if (!dev_priv->csr.dmc_payload)
> > -		intel_csr_runtime_pm_get(dev_priv);
> > +	if (!i915->csr.dmc_payload)
> > +		intel_csr_runtime_pm_get(i915);
> >   }
> >   /**
> >    * intel_csr_ucode_fini() - unload the CSR firmware.
> > - * @dev_priv: i915 drm device.
> > + * @i915: i915 drm device.
> >    *
> >    * Firmmware unloading includes freeing the internal memory and reset the
> >    * firmware loading status.
> >    */
> > -void intel_csr_ucode_fini(struct drm_i915_private *dev_priv)
> > +void intel_csr_ucode_fini(struct drm_i915_private *i915)
> >   {
> > -	if (!HAS_CSR(dev_priv))
> > +	if (!HAS_CSR(i915))
> >   		return;
> > -	intel_csr_ucode_suspend(dev_priv);
> > -	WARN_ON(dev_priv->csr.wakeref);
> > +	intel_csr_ucode_suspend(i915);
> > +	WARN_ON(i915->csr.wakeref);
> > -	kfree(dev_priv->csr.dmc_payload);
> > +	kfree(i915->csr.dmc_payload);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c
> > index 7925a176f900..eb61731f1abc 100644
> > --- a/drivers/gpu/drm/i915/intel_ddi.c
> > +++ b/drivers/gpu/drm/i915/intel_ddi.c
> > @@ -586,9 +586,9 @@ static const struct icl_mg_phy_ddi_buf_trans icl_mg_phy_ddi_translations[] = {
> >   };
> >   static const struct ddi_buf_trans *
> > -bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> > +bdw_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (dev_priv->vbt.edp.low_vswing) {
> > +	if (i915->vbt.edp.low_vswing) {
> >   		*n_entries = ARRAY_SIZE(bdw_ddi_translations_edp);
> >   		return bdw_ddi_translations_edp;
> >   	} else {
> > @@ -598,12 +598,12 @@ bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> >   }
> >   static const struct ddi_buf_trans *
> > -skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> > +skl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (IS_SKL_ULX(dev_priv)) {
> > +	if (IS_SKL_ULX(i915)) {
> >   		*n_entries = ARRAY_SIZE(skl_y_ddi_translations_dp);
> >   		return skl_y_ddi_translations_dp;
> > -	} else if (IS_SKL_ULT(dev_priv)) {
> > +	} else if (IS_SKL_ULT(i915)) {
> >   		*n_entries = ARRAY_SIZE(skl_u_ddi_translations_dp);
> >   		return skl_u_ddi_translations_dp;
> >   	} else {
> > @@ -613,12 +613,12 @@ skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> >   }
> >   static const struct ddi_buf_trans *
> > -kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> > +kbl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (IS_KBL_ULX(dev_priv) || IS_CFL_ULX(dev_priv)) {
> > +	if (IS_KBL_ULX(i915) || IS_CFL_ULX(i915)) {
> >   		*n_entries = ARRAY_SIZE(kbl_y_ddi_translations_dp);
> >   		return kbl_y_ddi_translations_dp;
> > -	} else if (IS_KBL_ULT(dev_priv) || IS_CFL_ULT(dev_priv)) {
> > +	} else if (IS_KBL_ULT(i915) || IS_CFL_ULT(i915)) {
> >   		*n_entries = ARRAY_SIZE(kbl_u_ddi_translations_dp);
> >   		return kbl_u_ddi_translations_dp;
> >   	} else {
> > @@ -628,15 +628,15 @@ kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> >   }
> >   static const struct ddi_buf_trans *
> > -skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> > +skl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (dev_priv->vbt.edp.low_vswing) {
> > -		if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) ||
> > -		    IS_CFL_ULX(dev_priv)) {
> > +	if (i915->vbt.edp.low_vswing) {
> > +		if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) ||
> > +		    IS_CFL_ULX(i915)) {
> >   			*n_entries = ARRAY_SIZE(skl_y_ddi_translations_edp);
> >   			return skl_y_ddi_translations_edp;
> > -		} else if (IS_SKL_ULT(dev_priv) || IS_KBL_ULT(dev_priv) ||
> > -			   IS_CFL_ULT(dev_priv)) {
> > +		} else if (IS_SKL_ULT(i915) || IS_KBL_ULT(i915) ||
> > +			   IS_CFL_ULT(i915)) {
> >   			*n_entries = ARRAY_SIZE(skl_u_ddi_translations_edp);
> >   			return skl_u_ddi_translations_edp;
> >   		} else {
> > @@ -645,17 +645,17 @@ skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> >   		}
> >   	}
> > -	if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> > -		return kbl_get_buf_trans_dp(dev_priv, n_entries);
> > +	if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915))
> > +		return kbl_get_buf_trans_dp(i915, n_entries);
> >   	else
> > -		return skl_get_buf_trans_dp(dev_priv, n_entries);
> > +		return skl_get_buf_trans_dp(i915, n_entries);
> >   }
> >   static const struct ddi_buf_trans *
> > -skl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> > +skl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) ||
> > -	    IS_CFL_ULX(dev_priv)) {
> > +	if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) ||
> > +	    IS_CFL_ULX(i915)) {
> >   		*n_entries = ARRAY_SIZE(skl_y_ddi_translations_hdmi);
> >   		return skl_y_ddi_translations_hdmi;
> >   	} else {
> > @@ -674,23 +674,23 @@ static int skl_buf_trans_num_entries(enum port port, int n_entries)
> >   }
> >   static const struct ddi_buf_trans *
> > -intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv,
> > +intel_ddi_get_buf_trans_dp(struct drm_i915_private *i915,
> >   			   enum port port, int *n_entries)
> >   {
> > -	if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> > +	if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> >   		const struct ddi_buf_trans *ddi_translations =
> > -			kbl_get_buf_trans_dp(dev_priv, n_entries);
> > +			kbl_get_buf_trans_dp(i915, n_entries);
> >   		*n_entries = skl_buf_trans_num_entries(port, *n_entries);
> >   		return ddi_translations;
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)) {
> >   		const struct ddi_buf_trans *ddi_translations =
> > -			skl_get_buf_trans_dp(dev_priv, n_entries);
> > +			skl_get_buf_trans_dp(i915, n_entries);
> >   		*n_entries = skl_buf_trans_num_entries(port, *n_entries);
> >   		return ddi_translations;
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > +	} else if (IS_BROADWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(bdw_ddi_translations_dp);
> >   		return  bdw_ddi_translations_dp;
> > -	} else if (IS_HASWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(hsw_ddi_translations_dp);
> >   		return hsw_ddi_translations_dp;
> >   	}
> > @@ -700,17 +700,17 @@ intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv,
> >   }
> >   static const struct ddi_buf_trans *
> > -intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv,
> > +intel_ddi_get_buf_trans_edp(struct drm_i915_private *i915,
> >   			    enum port port, int *n_entries)
> >   {
> > -	if (IS_GEN9_BC(dev_priv)) {
> > +	if (IS_GEN9_BC(i915)) {
> >   		const struct ddi_buf_trans *ddi_translations =
> > -			skl_get_buf_trans_edp(dev_priv, n_entries);
> > +			skl_get_buf_trans_edp(i915, n_entries);
> >   		*n_entries = skl_buf_trans_num_entries(port, *n_entries);
> >   		return ddi_translations;
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		return bdw_get_buf_trans_edp(dev_priv, n_entries);
> > -	} else if (IS_HASWELL(dev_priv)) {
> > +	} else if (IS_BROADWELL(i915)) {
> > +		return bdw_get_buf_trans_edp(i915, n_entries);
> > +	} else if (IS_HASWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(hsw_ddi_translations_dp);
> >   		return hsw_ddi_translations_dp;
> >   	}
> > @@ -720,13 +720,13 @@ intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv,
> >   }
> >   static const struct ddi_buf_trans *
> > -intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv,
> > +intel_ddi_get_buf_trans_fdi(struct drm_i915_private *i915,
> >   			    int *n_entries)
> >   {
> > -	if (IS_BROADWELL(dev_priv)) {
> > +	if (IS_BROADWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(bdw_ddi_translations_fdi);
> >   		return bdw_ddi_translations_fdi;
> > -	} else if (IS_HASWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(hsw_ddi_translations_fdi);
> >   		return hsw_ddi_translations_fdi;
> >   	}
> > @@ -736,15 +736,15 @@ intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv,
> >   }
> >   static const struct ddi_buf_trans *
> > -intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv,
> > +intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *i915,
> >   			     int *n_entries)
> >   {
> > -	if (IS_GEN9_BC(dev_priv)) {
> > -		return skl_get_buf_trans_hdmi(dev_priv, n_entries);
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > +	if (IS_GEN9_BC(i915)) {
> > +		return skl_get_buf_trans_hdmi(i915, n_entries);
> > +	} else if (IS_BROADWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(bdw_ddi_translations_hdmi);
> >   		return bdw_ddi_translations_hdmi;
> > -	} else if (IS_HASWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915)) {
> >   		*n_entries = ARRAY_SIZE(hsw_ddi_translations_hdmi);
> >   		return hsw_ddi_translations_hdmi;
> >   	}
> > @@ -754,32 +754,32 @@ intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv,
> >   }
> >   static const struct bxt_ddi_buf_trans *
> > -bxt_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> > +bxt_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> >   {
> >   	*n_entries = ARRAY_SIZE(bxt_ddi_translations_dp);
> >   	return bxt_ddi_translations_dp;
> >   }
> >   static const struct bxt_ddi_buf_trans *
> > -bxt_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> > +bxt_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> >   {
> > -	if (dev_priv->vbt.edp.low_vswing) {
> > +	if (i915->vbt.edp.low_vswing) {
> >   		*n_entries = ARRAY_SIZE(bxt_ddi_translations_edp);
> >   		return bxt_ddi_translations_edp;
> >   	}
> > -	return bxt_get_buf_trans_dp(dev_priv, n_entries);
> > +	return bxt_get_buf_trans_dp(i915, n_entries);
> >   }
> >   static const struct bxt_ddi_buf_trans *
> > -bxt_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> > +bxt_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> >   {
> >   	*n_entries = ARRAY_SIZE(bxt_ddi_translations_hdmi);
> >   	return bxt_ddi_translations_hdmi;
> >   }
> >   static const struct cnl_ddi_buf_trans *
> > -cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> > +cnl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries)
> >   {
> >   	u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
> > @@ -800,7 +800,7 @@ cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries)
> >   }
> >   static const struct cnl_ddi_buf_trans *
> > -cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> > +cnl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries)
> >   {
> >   	u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
> > @@ -821,11 +821,11 @@ cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries)
> >   }
> >   static const struct cnl_ddi_buf_trans *
> > -cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> > +cnl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries)
> >   {
> >   	u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
> > -	if (dev_priv->vbt.edp.low_vswing) {
> > +	if (i915->vbt.edp.low_vswing) {
> >   		if (voltage == VOLTAGE_INFO_0_85V) {
> >   			*n_entries = ARRAY_SIZE(cnl_ddi_translations_edp_0_85V);
> >   			return cnl_ddi_translations_edp_0_85V;
> > @@ -841,12 +841,12 @@ cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries)
> >   		}
> >   		return NULL;
> >   	} else {
> > -		return cnl_get_buf_trans_dp(dev_priv, n_entries);
> > +		return cnl_get_buf_trans_dp(i915, n_entries);
> >   	}
> >   }
> >   static const struct cnl_ddi_buf_trans *
> > -icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> > +icl_get_combo_buf_trans(struct drm_i915_private *i915, enum port port,
> >   			int type, int rate, int *n_entries)
> >   {
> >   	if (type == INTEL_OUTPUT_HDMI) {
> > @@ -855,7 +855,7 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> >   	} else if (rate > 540000 && type == INTEL_OUTPUT_EDP) {
> >   		*n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr3);
> >   		return icl_combo_phy_ddi_translations_edp_hbr3;
> > -	} else if (type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.low_vswing) {
> > +	} else if (type == INTEL_OUTPUT_EDP && i915->vbt.edp.low_vswing) {
> >   		*n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr2);
> >   		return icl_combo_phy_ddi_translations_edp_hbr2;
> >   	}
> > @@ -864,33 +864,33 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port,
> >   	return icl_combo_phy_ddi_translations_dp_hbr2;
> >   }
> > -static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port port)
> > +static int intel_ddi_hdmi_level(struct drm_i915_private *i915, enum port port)
> >   {
> >   	int n_entries, level, default_entry;
> > -	level = dev_priv->vbt.ddi_port_info[port].hdmi_level_shift;
> > +	level = i915->vbt.ddi_port_info[port].hdmi_level_shift;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		if (intel_port_is_combophy(dev_priv, port))
> > -			icl_get_combo_buf_trans(dev_priv, port, INTEL_OUTPUT_HDMI,
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		if (intel_port_is_combophy(i915, port))
> > +			icl_get_combo_buf_trans(i915, port, INTEL_OUTPUT_HDMI,
> >   						0, &n_entries);
> >   		else
> >   			n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations);
> >   		default_entry = n_entries - 1;
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > -		cnl_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	} else if (IS_CANNONLAKE(i915)) {
> > +		cnl_get_buf_trans_hdmi(i915, &n_entries);
> >   		default_entry = n_entries - 1;
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > -		bxt_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	} else if (IS_GEN9_LP(i915)) {
> > +		bxt_get_buf_trans_hdmi(i915, &n_entries);
> >   		default_entry = n_entries - 1;
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > -		intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	} else if (IS_GEN9_BC(i915)) {
> > +		intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> >   		default_entry = 8;
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	} else if (IS_BROADWELL(i915)) {
> > +		intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> >   		default_entry = 7;
> > -	} else if (IS_HASWELL(dev_priv)) {
> > -		intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	} else if (IS_HASWELL(i915)) {
> > +		intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> >   		default_entry = 6;
> >   	} else {
> >   		WARN(1, "ddi translation table missing\n");
> > @@ -917,25 +917,25 @@ static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port por
> >   static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder,
> >   					 const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 iboost_bit = 0;
> >   	int i, n_entries;
> >   	enum port port = encoder->port;
> >   	const struct ddi_buf_trans *ddi_translations;
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG))
> > -		ddi_translations = intel_ddi_get_buf_trans_fdi(dev_priv,
> > +		ddi_translations = intel_ddi_get_buf_trans_fdi(i915,
> >   							       &n_entries);
> >   	else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_EDP))
> > -		ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port,
> > +		ddi_translations = intel_ddi_get_buf_trans_edp(i915, port,
> >   							       &n_entries);
> >   	else
> > -		ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port,
> > +		ddi_translations = intel_ddi_get_buf_trans_dp(i915, port,
> >   							      &n_entries);
> >   	/* If we're boosting the current, set bit 31 of trans1 */
> > -	if (IS_GEN9_BC(dev_priv) &&
> > -	    dev_priv->vbt.ddi_port_info[port].dp_boost_level)
> > +	if (IS_GEN9_BC(i915) &&
> > +	    i915->vbt.ddi_port_info[port].dp_boost_level)
> >   		iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE;
> >   	for (i = 0; i < n_entries; i++) {
> > @@ -954,13 +954,13 @@ static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder,
> >   static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> >   					   int level)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 iboost_bit = 0;
> >   	int n_entries;
> >   	enum port port = encoder->port;
> >   	const struct ddi_buf_trans *ddi_translations;
> > -	ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +	ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> >   	if (WARN_ON_ONCE(!ddi_translations))
> >   		return;
> > @@ -968,8 +968,8 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> >   		level = n_entries - 1;
> >   	/* If we're boosting the current, set bit 31 of trans1 */
> > -	if (IS_GEN9_BC(dev_priv) &&
> > -	    dev_priv->vbt.ddi_port_info[port].hdmi_boost_level)
> > +	if (IS_GEN9_BC(i915) &&
> > +	    i915->vbt.ddi_port_info[port].hdmi_boost_level)
> >   		iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE;
> >   	/* Entry 9 is for HDMI: */
> > @@ -979,7 +979,7 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder,
> >   		   ddi_translations[level].trans2);
> >   }
> > -static void intel_wait_ddi_buf_idle(struct drm_i915_private *dev_priv,
> > +static void intel_wait_ddi_buf_idle(struct drm_i915_private *i915,
> >   				    enum port port)
> >   {
> >   	i915_reg_t reg = DDI_BUF_CTL(port);
> > @@ -1064,7 +1064,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> >   			const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_encoder *encoder;
> >   	u32 temp, i, rx_ctl_val, ddi_pll_sel;
> > @@ -1085,7 +1085,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> >   				  FDI_RX_TP1_TO_TP2_48 | FDI_RX_FDI_DELAY_90);
> >   	/* Enable the PCH Receiver FDI PLL */
> > -	rx_ctl_val = dev_priv->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE |
> > +	rx_ctl_val = i915->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE |
> >   		     FDI_RX_PLL_ENABLE |
> >   		     FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
> >   	I915_WRITE(FDI_RX_CTL(PIPE_A), rx_ctl_val);
> > @@ -1174,7 +1174,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc,
> >   		I915_WRITE(DP_TP_CTL(PORT_E), temp);
> >   		POSTING_READ(DP_TP_CTL(PORT_E));
> > -		intel_wait_ddi_buf_idle(dev_priv, PORT_E);
> > +		intel_wait_ddi_buf_idle(i915, PORT_E);
> >   		/* Reset FDI_RX_MISC pwrdn lanes */
> >   		temp = I915_READ(FDI_RX_MISC(PIPE_A));
> > @@ -1223,7 +1223,7 @@ intel_ddi_get_crtc_encoder(struct intel_crtc *crtc)
> >   	return ret;
> >   }
> > -static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv,
> > +static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *i915,
> >   				   i915_reg_t reg)
> >   {
> >   	int refclk;
> > @@ -1238,7 +1238,7 @@ static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv,
> >   		 * non-SSC for non-ULT HSW. Check FUSE_STRAP3
> >   		 * for the non-SSC reference frequency.
> >   		 */
> > -		if (IS_HASWELL(dev_priv) && !IS_HSW_ULT(dev_priv)) {
> > +		if (IS_HASWELL(i915) && !IS_HSW_ULT(i915)) {
> >   			if (I915_READ(FUSE_STRAP3) & HSW_REF_CLK_SELECT)
> >   				refclk = 24;
> >   			else
> > @@ -1325,7 +1325,7 @@ static int skl_calc_wrpll_link(const struct intel_dpll_hw_state *pll_state)
> >   	return dco_freq / (p0 * p1 * p2 * 5);
> >   }
> > -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> > +int cnl_calc_wrpll_link(struct drm_i915_private *i915,
> >   			struct intel_dpll_hw_state *pll_state)
> >   {
> >   	u32 p0, p1, p2, dco_freq, ref_clock;
> > @@ -1367,7 +1367,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> >   		break;
> >   	}
> > -	ref_clock = cnl_hdmi_pll_ref_clock(dev_priv);
> > +	ref_clock = cnl_hdmi_pll_ref_clock(i915);
> >   	dco_freq = (pll_state->cfgcr0 & DPLL_CFGCR0_DCO_INTEGER_MASK)
> >   		* ref_clock;
> > @@ -1381,7 +1381,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> >   	return dco_freq / (p0 * p1 * p2 * 5);
> >   }
> > -static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv,
> > +static int icl_calc_tbt_pll_link(struct drm_i915_private *i915,
> >   				 enum port port)
> >   {
> >   	u32 val = I915_READ(DDI_CLK_SEL(port)) & DDI_CLK_SEL_MASK;
> > @@ -1403,13 +1403,13 @@ static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int icl_calc_mg_pll_link(struct drm_i915_private *dev_priv,
> > +static int icl_calc_mg_pll_link(struct drm_i915_private *i915,
> >   				const struct intel_dpll_hw_state *pll_state)
> >   {
> >   	u32 m1, m2_int, m2_frac, div1, div2, ref_clock;
> >   	u64 tmp;
> > -	ref_clock = dev_priv->cdclk.hw.ref;
> > +	ref_clock = i915->cdclk.hw.ref;
> >   	m1 = pll_state->mg_pll_div1 & MG_PLL_DIV1_FBPREDIV_MASK;
> >   	m2_int = pll_state->mg_pll_div0 & MG_PLL_DIV0_FBDIV_INT_MASK;
> > @@ -1483,21 +1483,21 @@ static void ddi_dotclock_get(struct intel_crtc_state *pipe_config)
> >   static void icl_ddi_clock_get(struct intel_encoder *encoder,
> >   			      struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state;
> >   	enum port port = encoder->port;
> >   	int link_clock;
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > -		link_clock = cnl_calc_wrpll_link(dev_priv, pll_state);
> > +	if (intel_port_is_combophy(i915, port)) {
> > +		link_clock = cnl_calc_wrpll_link(i915, pll_state);
> >   	} else {
> > -		enum intel_dpll_id pll_id = intel_get_shared_dpll_id(dev_priv,
> > +		enum intel_dpll_id pll_id = intel_get_shared_dpll_id(i915,
> >   						pipe_config->shared_dpll);
> >   		if (pll_id == DPLL_ID_ICL_TBTPLL)
> > -			link_clock = icl_calc_tbt_pll_link(dev_priv, port);
> > +			link_clock = icl_calc_tbt_pll_link(i915, port);
> >   		else
> > -			link_clock = icl_calc_mg_pll_link(dev_priv, pll_state);
> > +			link_clock = icl_calc_mg_pll_link(i915, pll_state);
> >   	}
> >   	pipe_config->port_clock = link_clock;
> > @@ -1508,12 +1508,12 @@ static void icl_ddi_clock_get(struct intel_encoder *encoder,
> >   static void cnl_ddi_clock_get(struct intel_encoder *encoder,
> >   			      struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state;
> >   	int link_clock;
> >   	if (pll_state->cfgcr0 & DPLL_CFGCR0_HDMI_MODE) {
> > -		link_clock = cnl_calc_wrpll_link(dev_priv, pll_state);
> > +		link_clock = cnl_calc_wrpll_link(i915, pll_state);
> >   	} else {
> >   		link_clock = pll_state->cfgcr0 & DPLL_CFGCR0_LINK_RATE_MASK;
> > @@ -1604,7 +1604,7 @@ static void skl_ddi_clock_get(struct intel_encoder *encoder,
> >   static void hsw_ddi_clock_get(struct intel_encoder *encoder,
> >   			      struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	int link_clock = 0;
> >   	u32 val, pll;
> > @@ -1620,10 +1620,10 @@ static void hsw_ddi_clock_get(struct intel_encoder *encoder,
> >   		link_clock = 270000;
> >   		break;
> >   	case PORT_CLK_SEL_WRPLL1:
> > -		link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(0));
> > +		link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(0));
> >   		break;
> >   	case PORT_CLK_SEL_WRPLL2:
> > -		link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(1));
> > +		link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(1));
> >   		break;
> >   	case PORT_CLK_SEL_SPLL:
> >   		pll = I915_READ(SPLL_CTL) & SPLL_FREQ_MASK;
> > @@ -1675,24 +1675,24 @@ static void bxt_ddi_clock_get(struct intel_encoder *encoder,
> >   static void intel_ddi_clock_get(struct intel_encoder *encoder,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_ddi_clock_get(encoder, pipe_config);
> > -	else if (IS_CANNONLAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915))
> >   		cnl_ddi_clock_get(encoder, pipe_config);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		bxt_ddi_clock_get(encoder, pipe_config);
> > -	else if (IS_GEN9_BC(dev_priv))
> > +	else if (IS_GEN9_BC(i915))
> >   		skl_ddi_clock_get(encoder, pipe_config);
> > -	else if (INTEL_GEN(dev_priv) <= 8)
> > +	else if (INTEL_GEN(i915) <= 8)
> >   		hsw_ddi_clock_get(encoder, pipe_config);
> >   }
> >   void intel_ddi_set_pipe_settings(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 temp;
> > @@ -1746,7 +1746,7 @@ void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state,
> >   				    bool state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 temp;
> > @@ -1762,7 +1762,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	enum port port = encoder->port;
> > @@ -1845,7 +1845,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state)
> >   void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	i915_reg_t reg = TRANS_DDI_FUNC_CTL(cpu_transcoder);
> >   	u32 val = I915_READ(reg);
> > @@ -1854,7 +1854,7 @@ void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state
> >   	val |= TRANS_DDI_PORT_NONE;
> >   	I915_WRITE(reg, val);
> > -	if (dev_priv->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME &&
> > +	if (i915->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME &&
> >   	    intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> >   		DRM_DEBUG_KMS("Quirk Increase DDI disabled time\n");
> >   		/* Quirk time at 100ms for reliable operation */
> > @@ -1866,13 +1866,13 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> >   				     bool enable)
> >   {
> >   	struct drm_device *dev = intel_encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	intel_wakeref_t wakeref;
> >   	enum pipe pipe = 0;
> >   	int ret = 0;
> >   	u32 tmp;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     intel_encoder->power_domain);
> >   	if (WARN_ON(!wakeref))
> >   		return -ENXIO;
> > @@ -1889,14 +1889,14 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> >   		tmp &= ~TRANS_DDI_HDCP_SIGNALLING;
> >   	I915_WRITE(TRANS_DDI_FUNC_CTL(pipe), tmp);
> >   out:
> > -	intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, intel_encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> >   bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> >   {
> >   	struct drm_device *dev = intel_connector->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_encoder *encoder = intel_connector->encoder;
> >   	int type = intel_connector->base.connector_type;
> >   	enum port port = encoder->port;
> > @@ -1906,7 +1906,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> >   	u32 tmp;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1916,7 +1916,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> >   		goto out;
> >   	}
> > -	if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A)
> > +	if (HAS_TRANSCODER_EDP(i915) && port == PORT_A)
> >   		cpu_transcoder = TRANSCODER_EDP;
> >   	else
> >   		cpu_transcoder = (enum transcoder) pipe;
> > @@ -1950,7 +1950,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector)
> >   	}
> >   out:
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -1959,7 +1959,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   					u8 *pipe_mask, bool *is_dp_mst)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum port port = encoder->port;
> >   	intel_wakeref_t wakeref;
> >   	enum pipe p;
> > @@ -1969,7 +1969,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   	*pipe_mask = 0;
> >   	*is_dp_mst = false;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return;
> > @@ -1978,7 +1978,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   	if (!(tmp & DDI_BUF_CTL_ENABLE))
> >   		goto out;
> > -	if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A) {
> > +	if (HAS_TRANSCODER_EDP(i915) && port == PORT_A) {
> >   		tmp = I915_READ(TRANS_DDI_FUNC_CTL(TRANSCODER_EDP));
> >   		switch (tmp & TRANS_DDI_EDP_INPUT_MASK) {
> > @@ -2001,7 +2001,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   	}
> >   	mst_pipe_mask = 0;
> > -	for_each_pipe(dev_priv, p) {
> > +	for_each_pipe(i915, p) {
> >   		enum transcoder cpu_transcoder = (enum transcoder)p;
> >   		tmp = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder));
> > @@ -2033,7 +2033,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   		*is_dp_mst = mst_pipe_mask;
> >   out:
> > -	if (*pipe_mask && IS_GEN9_LP(dev_priv)) {
> > +	if (*pipe_mask && IS_GEN9_LP(i915)) {
> >   		tmp = I915_READ(BXT_PHY_CTL(port));
> >   		if ((tmp & (BXT_PHY_CMNLANE_POWERDOWN_ACK |
> >   			    BXT_PHY_LANE_POWERDOWN_ACK |
> > @@ -2042,7 +2042,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder,
> >   				  "(PHY_CTL %08x)\n", port_name(port), tmp);
> >   	}
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   }
> >   bool intel_ddi_get_hw_state(struct intel_encoder *encoder,
> > @@ -2083,7 +2083,7 @@ intel_ddi_main_link_aux_domain(struct intel_digital_port *dig_port)
> >   static void intel_ddi_get_power_domains(struct intel_encoder *encoder,
> >   					struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port;
> >   	/*
> > @@ -2095,29 +2095,29 @@ static void intel_ddi_get_power_domains(struct intel_encoder *encoder,
> >   		return;
> >   	dig_port = enc_to_dig_port(&encoder->base);
> > -	intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> > +	intel_display_power_get(i915, dig_port->ddi_io_power_domain);
> >   	/*
> >   	 * AUX power is only needed for (e)DP mode, and for HDMI mode on TC
> >   	 * ports.
> >   	 */
> >   	if (intel_crtc_has_dp_encoder(crtc_state) ||
> > -	    intel_port_is_tc(dev_priv, encoder->port))
> > -		intel_display_power_get(dev_priv,
> > +	    intel_port_is_tc(i915, encoder->port))
> > +		intel_display_power_get(i915,
> >   					intel_ddi_main_link_aux_domain(dig_port));
> >   	/*
> >   	 * VDSC power is needed when DSC is enabled
> >   	 */
> >   	if (crtc_state->dsc_params.compression_enable)
> > -		intel_display_power_get(dev_priv,
> > +		intel_display_power_get(i915,
> >   					intel_dsc_power_domain(crtc_state));
> >   }
> >   void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc);
> >   	enum port port = encoder->port;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> > @@ -2129,7 +2129,7 @@ void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state)
> >   void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	if (cpu_transcoder != TRANSCODER_EDP)
> > @@ -2137,7 +2137,7 @@ void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state)
> >   			   TRANS_CLK_SEL_DISABLED);
> >   }
> > -static void _skl_ddi_set_iboost(struct drm_i915_private *dev_priv,
> > +static void _skl_ddi_set_iboost(struct drm_i915_private *i915,
> >   				enum port port, u8 iboost)
> >   {
> >   	u32 tmp;
> > @@ -2155,25 +2155,25 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder,
> >   			       int level, enum intel_output_type type)
> >   {
> >   	struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	u8 iboost;
> >   	if (type == INTEL_OUTPUT_HDMI)
> > -		iboost = dev_priv->vbt.ddi_port_info[port].hdmi_boost_level;
> > +		iboost = i915->vbt.ddi_port_info[port].hdmi_boost_level;
> >   	else
> > -		iboost = dev_priv->vbt.ddi_port_info[port].dp_boost_level;
> > +		iboost = i915->vbt.ddi_port_info[port].dp_boost_level;
> >   	if (iboost == 0) {
> >   		const struct ddi_buf_trans *ddi_translations;
> >   		int n_entries;
> >   		if (type == INTEL_OUTPUT_HDMI)
> > -			ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +			ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries);
> >   		else if (type == INTEL_OUTPUT_EDP)
> > -			ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries);
> > +			ddi_translations = intel_ddi_get_buf_trans_edp(i915, port, &n_entries);
> >   		else
> > -			ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries);
> > +			ddi_translations = intel_ddi_get_buf_trans_dp(i915, port, &n_entries);
> >   		if (WARN_ON_ONCE(!ddi_translations))
> >   			return;
> > @@ -2189,33 +2189,33 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder,
> >   		return;
> >   	}
> > -	_skl_ddi_set_iboost(dev_priv, port, iboost);
> > +	_skl_ddi_set_iboost(i915, port, iboost);
> >   	if (port == PORT_A && intel_dig_port->max_lanes == 4)
> > -		_skl_ddi_set_iboost(dev_priv, PORT_E, iboost);
> > +		_skl_ddi_set_iboost(i915, PORT_E, iboost);
> >   }
> >   static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   				    int level, enum intel_output_type type)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	const struct bxt_ddi_buf_trans *ddi_translations;
> >   	enum port port = encoder->port;
> >   	int n_entries;
> >   	if (type == INTEL_OUTPUT_HDMI)
> > -		ddi_translations = bxt_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +		ddi_translations = bxt_get_buf_trans_hdmi(i915, &n_entries);
> >   	else if (type == INTEL_OUTPUT_EDP)
> > -		ddi_translations = bxt_get_buf_trans_edp(dev_priv, &n_entries);
> > +		ddi_translations = bxt_get_buf_trans_edp(i915, &n_entries);
> >   	else
> > -		ddi_translations = bxt_get_buf_trans_dp(dev_priv, &n_entries);
> > +		ddi_translations = bxt_get_buf_trans_dp(i915, &n_entries);
> >   	if (WARN_ON_ONCE(!ddi_translations))
> >   		return;
> >   	if (WARN_ON_ONCE(level >= n_entries))
> >   		level = n_entries - 1;
> > -	bxt_ddi_phy_set_signal_level(dev_priv, port,
> > +	bxt_ddi_phy_set_signal_level(i915, port,
> >   				     ddi_translations[level].margin,
> >   				     ddi_translations[level].scale,
> >   				     ddi_translations[level].enable,
> > @@ -2224,32 +2224,32 @@ static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   u8 intel_ddi_dp_voltage_max(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	enum port port = encoder->port;
> >   	int n_entries;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		if (intel_port_is_combophy(dev_priv, port))
> > -			icl_get_combo_buf_trans(dev_priv, port, encoder->type,
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		if (intel_port_is_combophy(i915, port))
> > +			icl_get_combo_buf_trans(i915, port, encoder->type,
> >   						intel_dp->link_rate, &n_entries);
> >   		else
> >   			n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations);
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_CANNONLAKE(i915)) {
> >   		if (encoder->type == INTEL_OUTPUT_EDP)
> > -			cnl_get_buf_trans_edp(dev_priv, &n_entries);
> > +			cnl_get_buf_trans_edp(i915, &n_entries);
> >   		else
> > -			cnl_get_buf_trans_dp(dev_priv, &n_entries);
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > +			cnl_get_buf_trans_dp(i915, &n_entries);
> > +	} else if (IS_GEN9_LP(i915)) {
> >   		if (encoder->type == INTEL_OUTPUT_EDP)
> > -			bxt_get_buf_trans_edp(dev_priv, &n_entries);
> > +			bxt_get_buf_trans_edp(i915, &n_entries);
> >   		else
> > -			bxt_get_buf_trans_dp(dev_priv, &n_entries);
> > +			bxt_get_buf_trans_dp(i915, &n_entries);
> >   	} else {
> >   		if (encoder->type == INTEL_OUTPUT_EDP)
> > -			intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries);
> > +			intel_ddi_get_buf_trans_edp(i915, port, &n_entries);
> >   		else
> > -			intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries);
> > +			intel_ddi_get_buf_trans_dp(i915, port, &n_entries);
> >   	}
> >   	if (WARN_ON(n_entries < 1))
> > @@ -2284,18 +2284,18 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder, u8 voltage_swing
> >   static void cnl_ddi_vswing_program(struct intel_encoder *encoder,
> >   				   int level, enum intel_output_type type)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	const struct cnl_ddi_buf_trans *ddi_translations;
> >   	enum port port = encoder->port;
> >   	int n_entries, ln;
> >   	u32 val;
> >   	if (type == INTEL_OUTPUT_HDMI)
> > -		ddi_translations = cnl_get_buf_trans_hdmi(dev_priv, &n_entries);
> > +		ddi_translations = cnl_get_buf_trans_hdmi(i915, &n_entries);
> >   	else if (type == INTEL_OUTPUT_EDP)
> > -		ddi_translations = cnl_get_buf_trans_edp(dev_priv, &n_entries);
> > +		ddi_translations = cnl_get_buf_trans_edp(i915, &n_entries);
> >   	else
> > -		ddi_translations = cnl_get_buf_trans_dp(dev_priv, &n_entries);
> > +		ddi_translations = cnl_get_buf_trans_dp(i915, &n_entries);
> >   	if (WARN_ON_ONCE(!ddi_translations))
> >   		return;
> > @@ -2348,7 +2348,7 @@ static void cnl_ddi_vswing_program(struct intel_encoder *encoder,
> >   static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   				    int level, enum intel_output_type type)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	int width, rate, ln;
> >   	u32 val;
> > @@ -2412,7 +2412,7 @@ static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   	I915_WRITE(CNL_PORT_TX_DW5_GRP(port), val);
> >   }
> > -static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv,
> > +static void icl_ddi_combo_vswing_program(struct drm_i915_private *i915,
> >   					u32 level, enum port port, int type,
> >   					int rate)
> >   {
> > @@ -2420,7 +2420,7 @@ static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv,
> >   	u32 n_entries, val;
> >   	int ln;
> > -	ddi_translations = icl_get_combo_buf_trans(dev_priv, port, type,
> > +	ddi_translations = icl_get_combo_buf_trans(i915, port, type,
> >   						   rate, &n_entries);
> >   	if (!ddi_translations)
> >   		return;
> > @@ -2472,7 +2472,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   					      u32 level,
> >   					      enum intel_output_type type)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	int width = 0;
> >   	int rate = 0;
> > @@ -2530,7 +2530,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   	I915_WRITE(ICL_PORT_TX_DW5_GRP(port), val);
> >   	/* 5. Program swing and de-emphasis */
> > -	icl_ddi_combo_vswing_program(dev_priv, level, port, type, rate);
> > +	icl_ddi_combo_vswing_program(i915, level, port, type, rate);
> >   	/* 6. Set training enable to trigger update */
> >   	val = I915_READ(ICL_PORT_TX_DW5_LN0(port));
> > @@ -2542,7 +2542,7 @@ static void icl_mg_phy_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   					   int link_clock,
> >   					   u32 level)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	const struct icl_mg_phy_ddi_buf_trans *ddi_translations;
> >   	u32 n_entries, val;
> > @@ -2662,10 +2662,10 @@ static void icl_ddi_vswing_sequence(struct intel_encoder *encoder,
> >   				    u32 level,
> >   				    enum intel_output_type type)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> > -	if (intel_port_is_combophy(dev_priv, port))
> > +	if (intel_port_is_combophy(i915, port))
> >   		icl_combo_phy_ddi_vswing_sequence(encoder, level, type);
> >   	else
> >   		icl_mg_phy_ddi_vswing_sequence(encoder, link_clock, level);
> > @@ -2698,14 +2698,14 @@ static u32 intel_ddi_dp_level(struct intel_dp *intel_dp)
> >   u32 bxt_signal_levels(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
> >   	struct intel_encoder *encoder = &dport->base;
> >   	int level = intel_ddi_dp_level(intel_dp);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_ddi_vswing_sequence(encoder, intel_dp->link_rate,
> >   					level, encoder->type);
> > -	else if (IS_CANNONLAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915))
> >   		cnl_ddi_vswing_sequence(encoder, level, encoder->type);
> >   	else
> >   		bxt_ddi_vswing_sequence(encoder, level, encoder->type);
> > @@ -2716,24 +2716,24 @@ u32 bxt_signal_levels(struct intel_dp *intel_dp)
> >   u32 ddi_signal_levels(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
> >   	struct intel_encoder *encoder = &dport->base;
> >   	int level = intel_ddi_dp_level(intel_dp);
> > -	if (IS_GEN9_BC(dev_priv))
> > +	if (IS_GEN9_BC(i915))
> >   		skl_ddi_set_iboost(encoder, level, encoder->type);
> >   	return DDI_BUF_TRANS_SELECT(level);
> >   }
> >   static inline
> > -u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv,
> > +u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *i915,
> >   			      enum port port)
> >   {
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > +	if (intel_port_is_combophy(i915, port)) {
> >   		return ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> > -	} else if (intel_port_is_tc(dev_priv, port)) {
> > -		enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	} else if (intel_port_is_tc(i915, port)) {
> > +		enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   		return ICL_DPCLKA_CFGCR0_TC_CLK_OFF(tc_port);
> >   	}
> > @@ -2744,47 +2744,47 @@ u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv,
> >   static void icl_map_plls_to_ports(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> >   	enum port port = encoder->port;
> >   	u32 val;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	val = I915_READ(DPCLKA_CFGCR0_ICL);
> > -	WARN_ON((val & icl_dpclka_cfgcr0_clk_off(dev_priv, port)) == 0);
> > +	WARN_ON((val & icl_dpclka_cfgcr0_clk_off(i915, port)) == 0);
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > +	if (intel_port_is_combophy(i915, port)) {
> >   		val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> >   		val |= DPCLKA_CFGCR0_DDI_CLK_SEL(pll->info->id, port);
> >   		I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> >   		POSTING_READ(DPCLKA_CFGCR0_ICL);
> >   	}
> > -	val &= ~icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> > +	val &= ~icl_dpclka_cfgcr0_clk_off(i915, port);
> >   	I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static void icl_unmap_plls_to_ports(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	u32 val;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	val = I915_READ(DPCLKA_CFGCR0_ICL);
> > -	val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> > +	val |= icl_dpclka_cfgcr0_clk_off(i915, port);
> >   	I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val;
> >   	enum port port;
> >   	u32 port_mask;
> > @@ -2821,7 +2821,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> >   		 * Sanity check that we haven't incorrectly registered another
> >   		 * encoder using any of the ports of this DSI encoder.
> >   		 */
> > -		for_each_intel_encoder(&dev_priv->drm, other_encoder) {
> > +		for_each_intel_encoder(&i915->drm, other_encoder) {
> >   			if (other_encoder == encoder)
> >   				continue;
> > @@ -2838,7 +2838,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> >   	val = I915_READ(DPCLKA_CFGCR0_ICL);
> >   	for_each_port_masked(port, port_mask) {
> >   		bool ddi_clk_ungated = !(val &
> > -					 icl_dpclka_cfgcr0_clk_off(dev_priv,
> > +					 icl_dpclka_cfgcr0_clk_off(i915,
> >   								   port));
> >   		if (ddi_clk_needed == ddi_clk_ungated)
> > @@ -2853,7 +2853,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> >   		DRM_NOTE("Port %c is disabled/in DSI mode with an ungated DDI clock, gate it\n",
> >   			 port_name(port));
> > -		val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port);
> > +		val |= icl_dpclka_cfgcr0_clk_off(i915, port);
> >   		I915_WRITE(DPCLKA_CFGCR0_ICL, val);
> >   	}
> >   }
> > @@ -2861,7 +2861,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder)
> >   static void intel_ddi_clk_select(struct intel_encoder *encoder,
> >   				 const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	u32 val;
> >   	const struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> > @@ -2869,13 +2869,13 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
> >   	if (WARN_ON(!pll))
> >   		return;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		if (!intel_port_is_combophy(dev_priv, port))
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		if (!intel_port_is_combophy(i915, port))
> >   			I915_WRITE(DDI_CLK_SEL(port),
> >   				   icl_pll_to_ddi_clk_sel(encoder, crtc_state));
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_CANNONLAKE(i915)) {
> >   		/* Configure DPCLKA_CFGCR0 to map the DPLL to the DDI. */
> >   		val = I915_READ(DPCLKA_CFGCR0);
> >   		val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> > @@ -2890,7 +2890,7 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
> >   		val = I915_READ(DPCLKA_CFGCR0);
> >   		val &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port);
> >   		I915_WRITE(DPCLKA_CFGCR0, val);
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > +	} else if (IS_GEN9_BC(i915)) {
> >   		/* DDI -> PLL mapping  */
> >   		val = I915_READ(DPLL_CTRL2);
> > @@ -2901,37 +2901,37 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder,
> >   		I915_WRITE(DPLL_CTRL2, val);
> > -	} else if (INTEL_GEN(dev_priv) < 9) {
> > +	} else if (INTEL_GEN(i915) < 9) {
> >   		I915_WRITE(PORT_CLK_SEL(port), hsw_pll_to_ddi_pll_sel(pll));
> >   	}
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static void intel_ddi_clk_disable(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		if (!intel_port_is_combophy(dev_priv, port))
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		if (!intel_port_is_combophy(i915, port))
> >   			I915_WRITE(DDI_CLK_SEL(port), DDI_CLK_SEL_NONE);
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_CANNONLAKE(i915)) {
> >   		I915_WRITE(DPCLKA_CFGCR0, I915_READ(DPCLKA_CFGCR0) |
> >   			   DPCLKA_CFGCR0_DDI_CLK_OFF(port));
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > +	} else if (IS_GEN9_BC(i915)) {
> >   		I915_WRITE(DPLL_CTRL2, I915_READ(DPLL_CTRL2) |
> >   			   DPLL_CTRL2_DDI_CLK_OFF(port));
> > -	} else if (INTEL_GEN(dev_priv) < 9) {
> > +	} else if (INTEL_GEN(i915) < 9) {
> >   		I915_WRITE(PORT_CLK_SEL(port), PORT_CLK_SEL_NONE);
> >   	}
> >   }
> >   static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	enum port port = dig_port->base.port;
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   	u32 val;
> >   	int ln;
> > @@ -2961,9 +2961,9 @@ static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port)
> >   static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	enum port port = dig_port->base.port;
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   	u32 val;
> >   	int ln;
> > @@ -2993,9 +2993,9 @@ static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port)
> >   static void icl_program_mg_dp_mode(struct intel_digital_port *intel_dig_port)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> >   	enum port port = intel_dig_port->base.port;
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   	u32 ln0, ln1, lane_info;
> >   	if (tc_port == PORT_TC_NONE || intel_dig_port->tc_type == TC_PORT_TBT)
> > @@ -3069,7 +3069,7 @@ static void intel_dp_sink_set_fec_ready(struct intel_dp *intel_dp,
> >   static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> >   				 const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	u32 val;
> > @@ -3080,7 +3080,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> >   	val |= DP_TP_CTL_FEC_ENABLE;
> >   	I915_WRITE(DP_TP_CTL(port), val);
> > -	if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port),
> > +	if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port),
> >   				    DP_TP_STATUS_FEC_ENABLE_LIVE,
> >   				    DP_TP_STATUS_FEC_ENABLE_LIVE,
> >   				    1))
> > @@ -3090,7 +3090,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder,
> >   static void intel_ddi_disable_fec_state(struct intel_encoder *encoder,
> >   					const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	u32 val;
> > @@ -3108,7 +3108,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
> >   				    const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	bool is_mst = intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DP_MST);
> > @@ -3123,26 +3123,26 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
> >   	intel_ddi_clk_select(encoder, crtc_state);
> > -	intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> > +	intel_display_power_get(i915, dig_port->ddi_io_power_domain);
> >   	icl_program_mg_dp_mode(dig_port);
> >   	icl_disable_phy_clock_gating(dig_port);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_ddi_vswing_sequence(encoder, crtc_state->port_clock,
> >   					level, encoder->type);
> > -	else if (IS_CANNONLAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915))
> >   		cnl_ddi_vswing_sequence(encoder, level, encoder->type);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		bxt_ddi_vswing_sequence(encoder, level, encoder->type);
> >   	else
> >   		intel_prepare_dp_ddi_buffers(encoder, crtc_state);
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > +	if (intel_port_is_combophy(i915, port)) {
> >   		bool lane_reversal =
> >   			dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL;
> > -		intel_combo_phy_power_up_lanes(dev_priv, port, false,
> > +		intel_combo_phy_power_up_lanes(i915, port, false,
> >   					       crtc_state->lane_count,
> >   					       lane_reversal);
> >   	}
> > @@ -3154,7 +3154,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder,
> >   					      true);
> >   	intel_dp_sink_set_fec_ready(intel_dp, crtc_state);
> >   	intel_dp_start_link_train(intel_dp);
> > -	if (port != PORT_A || INTEL_GEN(dev_priv) >= 9)
> > +	if (port != PORT_A || INTEL_GEN(i915) >= 9)
> >   		intel_dp_stop_link_train(intel_dp);
> >   	intel_ddi_enable_fec(encoder, crtc_state);
> > @@ -3173,32 +3173,32 @@ static void intel_ddi_pre_enable_hdmi(struct intel_encoder *encoder,
> >   {
> >   	struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> >   	struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> > -	int level = intel_ddi_hdmi_level(dev_priv, port);
> > +	int level = intel_ddi_hdmi_level(i915, port);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	intel_dp_dual_mode_set_tmds_output(intel_hdmi, true);
> >   	intel_ddi_clk_select(encoder, crtc_state);
> > -	intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain);
> > +	intel_display_power_get(i915, dig_port->ddi_io_power_domain);
> >   	icl_program_mg_dp_mode(dig_port);
> >   	icl_disable_phy_clock_gating(dig_port);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_ddi_vswing_sequence(encoder, crtc_state->port_clock,
> >   					level, INTEL_OUTPUT_HDMI);
> > -	else if (IS_CANNONLAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915))
> >   		cnl_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		bxt_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI);
> >   	else
> >   		intel_prepare_hdmi_ddi_buffers(encoder, level);
> >   	icl_enable_phy_clock_gating(dig_port);
> > -	if (IS_GEN9_BC(dev_priv))
> > +	if (IS_GEN9_BC(i915))
> >   		skl_ddi_set_iboost(encoder, level, INTEL_OUTPUT_HDMI);
> >   	intel_ddi_enable_pipe_clock(crtc_state);
> > @@ -3213,7 +3213,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/*
> > @@ -3231,10 +3231,10 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
> >   	WARN_ON(crtc_state->has_pch_encoder);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_map_plls_to_ports(encoder, crtc_state);
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> >   		intel_ddi_pre_enable_hdmi(encoder, crtc_state, conn_state);
> > @@ -3257,7 +3257,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder,
> >   static void intel_disable_ddi_buf(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	bool wait = false;
> >   	u32 val;
> > @@ -3278,14 +3278,14 @@ static void intel_disable_ddi_buf(struct intel_encoder *encoder,
> >   	intel_ddi_disable_fec_state(encoder, crtc_state);
> >   	if (wait)
> > -		intel_wait_ddi_buf_idle(dev_priv, port);
> > +		intel_wait_ddi_buf_idle(i915, port);
> >   }
> >   static void intel_ddi_post_disable_dp(struct intel_encoder *encoder,
> >   				      const struct intel_crtc_state *old_crtc_state,
> >   				      const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	struct intel_dp *intel_dp = &dig_port->dp;
> >   	bool is_mst = intel_crtc_has_type(old_crtc_state,
> > @@ -3305,7 +3305,7 @@ static void intel_ddi_post_disable_dp(struct intel_encoder *encoder,
> >   	intel_edp_panel_vdd_on(intel_dp);
> >   	intel_edp_panel_off(intel_dp);
> > -	intel_display_power_put_unchecked(dev_priv,
> > +	intel_display_power_put_unchecked(i915,
> >   					  dig_port->ddi_io_power_domain);
> >   	intel_ddi_clk_disable(encoder);
> > @@ -3315,7 +3315,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder,
> >   					const struct intel_crtc_state *old_crtc_state,
> >   					const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	struct intel_hdmi *intel_hdmi = &dig_port->hdmi;
> > @@ -3326,7 +3326,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder,
> >   	intel_disable_ddi_buf(encoder, old_crtc_state);
> > -	intel_display_power_put_unchecked(dev_priv,
> > +	intel_display_power_put_unchecked(i915,
> >   					  dig_port->ddi_io_power_domain);
> >   	intel_ddi_clk_disable(encoder);
> > @@ -3338,7 +3338,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *old_crtc_state,
> >   				   const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	/*
> >   	 * When called from DP MST code:
> > @@ -3360,7 +3360,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder,
> >   		intel_ddi_post_disable_dp(encoder,
> >   					  old_crtc_state, old_conn_state);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_unmap_plls_to_ports(encoder);
> >   }
> > @@ -3368,7 +3368,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *encoder,
> >   				const struct intel_crtc_state *old_crtc_state,
> >   				const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val;
> >   	/*
> > @@ -3402,11 +3402,11 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder,
> >   				const struct intel_crtc_state *crtc_state,
> >   				const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	enum port port = encoder->port;
> > -	if (port == PORT_A && INTEL_GEN(dev_priv) < 9)
> > +	if (port == PORT_A && INTEL_GEN(i915) < 9)
> >   		intel_dp_stop_link_train(intel_dp);
> >   	intel_edp_backlight_on(crtc_state, conn_state);
> > @@ -3419,7 +3419,7 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder,
> >   }
> >   static i915_reg_t
> > -gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv,
> > +gen9_chicken_trans_reg_by_port(struct drm_i915_private *i915,
> >   			       enum port port)
> >   {
> >   	static const i915_reg_t regs[] = {
> > @@ -3430,7 +3430,7 @@ gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv,
> >   		[PORT_E] = CHICKEN_TRANS_A,
> >   	};
> > -	WARN_ON(INTEL_GEN(dev_priv) < 9);
> > +	WARN_ON(INTEL_GEN(i915) < 9);
> >   	if (WARN_ON(port < PORT_A || port > PORT_E))
> >   		port = PORT_A;
> > @@ -3442,7 +3442,7 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	struct drm_connector *connector = conn_state->connector;
> >   	enum port port = encoder->port;
> > @@ -3454,14 +3454,14 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder,
> >   			  connector->base.id, connector->name);
> >   	/* Display WA #1143: skl,kbl,cfl */
> > -	if (IS_GEN9_BC(dev_priv)) {
> > +	if (IS_GEN9_BC(i915)) {
> >   		/*
> >   		 * For some reason these chicken bits have been
> >   		 * stuffed into a transcoder register, event though
> >   		 * the bits affect a specific DDI port rather than
> >   		 * a specific transcoder.
> >   		 */
> > -		i915_reg_t reg = gen9_chicken_trans_reg_by_port(dev_priv, port);
> > +		i915_reg_t reg = gen9_chicken_trans_reg_by_port(i915, port);
> >   		u32 val;
> >   		val = I915_READ(reg);
> > @@ -3595,9 +3595,9 @@ static void intel_ddi_set_fia_lane_count(struct intel_encoder *encoder,
> >   					 const struct intel_crtc_state *pipe_config,
> >   					 enum port port)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   	u32 val = I915_READ(PORT_TX_DFLEXDPMLE1);
> >   	bool lane_reversal = dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL;
> > @@ -3625,16 +3625,16 @@ intel_ddi_pre_pll_enable(struct intel_encoder *encoder,
> >   			 const struct intel_crtc_state *crtc_state,
> >   			 const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	enum port port = encoder->port;
> >   	if (intel_crtc_has_dp_encoder(crtc_state) ||
> > -	    intel_port_is_tc(dev_priv, encoder->port))
> > -		intel_display_power_get(dev_priv,
> > +	    intel_port_is_tc(i915, encoder->port))
> > +		intel_display_power_get(i915,
> >   					intel_ddi_main_link_aux_domain(dig_port));
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		bxt_ddi_phy_set_lane_optim_mask(encoder,
> >   						crtc_state->lane_lat_optim_mask);
> > @@ -3654,19 +3654,19 @@ intel_ddi_post_pll_disable(struct intel_encoder *encoder,
> >   			   const struct intel_crtc_state *crtc_state,
> >   			   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	if (intel_crtc_has_dp_encoder(crtc_state) ||
> > -	    intel_port_is_tc(dev_priv, encoder->port))
> > -		intel_display_power_put_unchecked(dev_priv,
> > +	    intel_port_is_tc(i915, encoder->port))
> > +		intel_display_power_put_unchecked(i915,
> >   						  intel_ddi_main_link_aux_domain(dig_port));
> >   }
> >   static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(intel_dig_port->base.base.dev);
> >   	enum port port = intel_dig_port->base.port;
> >   	u32 val;
> > @@ -3687,7 +3687,7 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> >   		POSTING_READ(DP_TP_CTL(port));
> >   		if (wait)
> > -			intel_wait_ddi_buf_idle(dev_priv, port);
> > +			intel_wait_ddi_buf_idle(i915, port);
> >   	}
> >   	val = DP_TP_CTL_ENABLE |
> > @@ -3709,32 +3709,32 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp)
> >   	udelay(600);
> >   }
> > -static bool intel_ddi_is_audio_enabled(struct drm_i915_private *dev_priv,
> > +static bool intel_ddi_is_audio_enabled(struct drm_i915_private *i915,
> >   				       enum transcoder cpu_transcoder)
> >   {
> >   	if (cpu_transcoder == TRANSCODER_EDP)
> >   		return false;
> > -	if (!intel_display_power_is_enabled(dev_priv, POWER_DOMAIN_AUDIO))
> > +	if (!intel_display_power_is_enabled(i915, POWER_DOMAIN_AUDIO))
> >   		return false;
> >   	return I915_READ(HSW_AUD_PIN_ELD_CP_VLD) &
> >   		AUDIO_OUTPUT_ENABLE(cpu_transcoder);
> >   }
> > -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv,
> > +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915,
> >   					 struct intel_crtc_state *crtc_state)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 11 && crtc_state->port_clock > 594000)
> > +	if (INTEL_GEN(i915) >= 11 && crtc_state->port_clock > 594000)
> >   		crtc_state->min_voltage_level = 1;
> > -	else if (IS_CANNONLAKE(dev_priv) && crtc_state->port_clock > 594000)
> > +	else if (IS_CANNONLAKE(i915) && crtc_state->port_clock > 594000)
> >   		crtc_state->min_voltage_level = 2;
> >   }
> >   void intel_ddi_get_config(struct intel_encoder *encoder,
> >   			  struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> >   	struct intel_digital_port *intel_dig_port;
> > @@ -3816,10 +3816,10 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> >   	}
> >   	pipe_config->has_audio =
> > -		intel_ddi_is_audio_enabled(dev_priv, cpu_transcoder);
> > +		intel_ddi_is_audio_enabled(i915, cpu_transcoder);
> > -	if (encoder->type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.bpp &&
> > -	    pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) {
> > +	if (encoder->type == INTEL_OUTPUT_EDP && i915->vbt.edp.bpp &&
> > +	    pipe_config->pipe_bpp > i915->vbt.edp.bpp) {
> >   		/*
> >   		 * This is a big fat ugly hack.
> >   		 *
> > @@ -3834,17 +3834,17 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> >   		 * load.
> >   		 */
> >   		DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n",
> > -			      pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp);
> > -		dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp;
> > +			      pipe_config->pipe_bpp, i915->vbt.edp.bpp);
> > +		i915->vbt.edp.bpp = pipe_config->pipe_bpp;
> >   	}
> >   	intel_ddi_clock_get(encoder, pipe_config);
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		pipe_config->lane_lat_optim_mask =
> >   			bxt_ddi_phy_get_lane_lat_optim_mask(encoder);
> > -	intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> > +	intel_ddi_compute_min_voltage_level(i915, pipe_config);
> >   	intel_hdmi_read_gcp_infoframe(encoder, pipe_config);
> > @@ -3885,11 +3885,11 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder,
> >   				    struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	int ret;
> > -	if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A)
> > +	if (HAS_TRANSCODER_EDP(i915) && port == PORT_A)
> >   		pipe_config->cpu_transcoder = TRANSCODER_EDP;
> >   	if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI))
> > @@ -3899,17 +3899,17 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder,
> >   	if (ret)
> >   		return ret;
> > -	if (IS_HASWELL(dev_priv) && crtc->pipe == PIPE_A &&
> > +	if (IS_HASWELL(i915) && crtc->pipe == PIPE_A &&
> >   	    pipe_config->cpu_transcoder == TRANSCODER_EDP)
> >   		pipe_config->pch_pfit.force_thru =
> >   			pipe_config->pch_pfit.enabled ||
> >   			pipe_config->crc_enabled;
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		pipe_config->lane_lat_optim_mask =
> >   			bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count);
> > -	intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> > +	intel_ddi_compute_min_voltage_level(i915, pipe_config);
> >   	return 0;
> >   }
> > @@ -4013,11 +4013,11 @@ static int modeset_pipe(struct drm_crtc *crtc,
> >   static int intel_hdmi_reset_link(struct intel_encoder *encoder,
> >   				 struct drm_modeset_acquire_ctx *ctx)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_hdmi *hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	struct intel_connector *connector = hdmi->attached_connector;
> >   	struct i2c_adapter *adapter =
> > -		intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> > +		intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
> >   	struct drm_connector_state *conn_state;
> >   	struct intel_crtc_state *crtc_state;
> >   	struct intel_crtc *crtc;
> > @@ -4027,7 +4027,7 @@ static int intel_hdmi_reset_link(struct intel_encoder *encoder,
> >   	if (!connector || connector->base.status != connector_status_connected)
> >   		return 0;
> > -	ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> > +	ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> >   			       ctx);
> >   	if (ret)
> >   		return ret;
> > @@ -4131,7 +4131,7 @@ intel_ddi_init_hdmi_connector(struct intel_digital_port *intel_dig_port)
> >   static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dport->base.base.dev);
> >   	if (dport->base.port != PORT_A)
> >   		return false;
> > @@ -4142,7 +4142,7 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> >   	/* Broxton/Geminilake: Bspec says that DDI_A_4_LANES is the only
> >   	 *                     supported configuration
> >   	 */
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		return true;
> >   	/* Cannonlake: Most of SKUs don't support DDI_E, and the only
> > @@ -4150,8 +4150,8 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> >   	 *             DDI_F what makes DDI_E useless. However for this
> >   	 *             case let's trust VBT info.
> >   	 */
> > -	if (IS_CANNONLAKE(dev_priv) &&
> > -	    !intel_bios_is_port_present(dev_priv, PORT_E))
> > +	if (IS_CANNONLAKE(i915) &&
> > +	    !intel_bios_is_port_present(i915, PORT_E))
> >   		return true;
> >   	return false;
> > @@ -4160,11 +4160,11 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport)
> >   static int
> >   intel_ddi_max_lanes(struct intel_digital_port *intel_dport)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dport->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dport->base.base.dev);
> >   	enum port port = intel_dport->base.port;
> >   	int max_lanes = 4;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return max_lanes;
> >   	if (port == PORT_A || port == PORT_E) {
> > @@ -4189,10 +4189,10 @@ intel_ddi_max_lanes(struct intel_digital_port *intel_dport)
> >   	return max_lanes;
> >   }
> > -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> > +void intel_ddi_init(struct drm_i915_private *i915, enum port port)
> >   {
> >   	struct ddi_vbt_port_info *port_info =
> > -		&dev_priv->vbt.ddi_port_info[port];
> > +		&i915->vbt.ddi_port_info[port];
> >   	struct intel_digital_port *intel_dig_port;
> >   	struct intel_encoder *intel_encoder;
> >   	struct drm_encoder *encoder;
> > @@ -4202,7 +4202,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> >   	init_hdmi = port_info->supports_dvi || port_info->supports_hdmi;
> >   	init_dp = port_info->supports_dp;
> > -	if (intel_bios_is_lspcon_present(dev_priv, port)) {
> > +	if (intel_bios_is_lspcon_present(i915, port)) {
> >   		/*
> >   		 * Lspcon device needs to be driven with DP connector
> >   		 * with special detection sequence. So make sure DP
> > @@ -4227,7 +4227,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> >   	intel_encoder = &intel_dig_port->base;
> >   	encoder = &intel_encoder->base;
> > -	drm_encoder_init(&dev_priv->drm, encoder, &intel_ddi_funcs,
> > +	drm_encoder_init(&i915->drm, encoder, &intel_ddi_funcs,
> >   			 DRM_MODE_ENCODER_TMDS, "DDI %c", port_name(port));
> >   	intel_encoder->hotplug = intel_ddi_hotplug;
> > @@ -4248,10 +4248,10 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> >   	intel_encoder->power_domain = intel_port_to_power_domain(port);
> >   	intel_encoder->port = port;
> >   	intel_encoder->cloneable = 0;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		intel_encoder->crtc_mask |= BIT(pipe);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		intel_dig_port->saved_port_bits = I915_READ(DDI_BUF_CTL(port)) &
> >   			DDI_BUF_PORT_REVERSAL;
> >   	else
> > @@ -4259,9 +4259,9 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> >   			(DDI_BUF_PORT_REVERSAL | DDI_A_4_LANES);
> >   	intel_dig_port->dp.output_reg = INVALID_MMIO_REG;
> >   	intel_dig_port->max_lanes = intel_ddi_max_lanes(intel_dig_port);
> > -	intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> > +	intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
> > -	intel_dig_port->tc_legacy_port = intel_port_is_tc(dev_priv, port) &&
> > +	intel_dig_port->tc_legacy_port = intel_port_is_tc(i915, port) &&
> >   					 !port_info->supports_typec_usb &&
> >   					 !port_info->supports_tbt;
> > @@ -4324,7 +4324,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port)
> >   	intel_infoframe_init(intel_dig_port);
> > -	if (intel_port_is_tc(dev_priv, port))
> > +	if (intel_port_is_tc(i915, port))
> >   		intel_digital_port_connected(intel_encoder);
> >   	return;
> > diff --git a/drivers/gpu/drm/i915/intel_ddi.h b/drivers/gpu/drm/i915/intel_ddi.h
> > index a08365da2643..50dd6fac4e3b 100644
> > --- a/drivers/gpu/drm/i915/intel_ddi.h
> > +++ b/drivers/gpu/drm/i915/intel_ddi.h
> > @@ -24,7 +24,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *intel_encoder,
> >   				const struct drm_connector_state *old_conn_state);
> >   void hsw_fdi_link_train(struct intel_crtc *crtc,
> >   			const struct intel_crtc_state *crtc_state);
> > -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port);
> > +void intel_ddi_init(struct drm_i915_private *i915, enum port port);
> >   bool intel_ddi_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe);
> >   void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state);
> >   void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state);
> > @@ -36,7 +36,7 @@ void intel_ddi_get_config(struct intel_encoder *encoder,
> >   			  struct intel_crtc_state *pipe_config);
> >   void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state,
> >   				    bool state);
> > -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv,
> > +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915,
> >   					 struct intel_crtc_state *crtc_state);
> >   u32 bxt_signal_levels(struct intel_dp *intel_dp);
> >   u32 ddi_signal_levels(struct intel_dp *intel_dp);
> > @@ -46,7 +46,7 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder,
> >   int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder,
> >   				     bool enable);
> >   void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder);
> > -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv,
> > +int cnl_calc_wrpll_link(struct drm_i915_private *i915,
> >   			struct intel_dpll_hw_state *state);
> >   #endif /* __INTEL_DDI_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_device_info.c b/drivers/gpu/drm/i915/intel_device_info.c
> > index 7135d8dc32a7..a670881b3f1e 100644
> > --- a/drivers/gpu/drm/i915/intel_device_info.c
> > +++ b/drivers/gpu/drm/i915/intel_device_info.c
> > @@ -182,15 +182,15 @@ static u16 compute_eu_total(const struct sseu_dev_info *sseu)
> >   	return total;
> >   }
> > -static void gen11_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void gen11_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	u8 s_en;
> >   	u32 ss_en, ss_en_mask;
> >   	u8 eu_en;
> >   	int s;
> > -	if (IS_ELKHARTLAKE(dev_priv)) {
> > +	if (IS_ELKHARTLAKE(i915)) {
> >   		sseu->max_slices = 1;
> >   		sseu->max_subslices = 4;
> >   		sseu->max_eus_per_subslice = 8;
> > @@ -227,9 +227,9 @@ static void gen11_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->has_eu_pg = 1;
> >   }
> > -static void gen10_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void gen10_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	const u32 fuse2 = I915_READ(GEN8_FUSE2);
> >   	int s, ss;
> >   	const int eu_mask = 0xff;
> > @@ -305,9 +305,9 @@ static void gen10_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->has_eu_pg = 1;
> >   }
> > -static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void cherryview_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	u32 fuse;
> >   	fuse = I915_READ(CHV_FUSE_GT);
> > @@ -359,10 +359,10 @@ static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->has_eu_pg = (sseu->eu_per_subslice > 2);
> >   }
> > -static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void gen9_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct intel_device_info *info = mkwrite_device_info(dev_priv);
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct intel_device_info *info = mkwrite_device_info(i915);
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	int s, ss;
> >   	u32 fuse2, eu_disable, subslice_mask;
> >   	const u8 eu_mask = 0xff;
> > @@ -371,8 +371,8 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->slice_mask = (fuse2 & GEN8_F2_S_ENA_MASK) >> GEN8_F2_S_ENA_SHIFT;
> >   	/* BXT has a single slice and at most 3 subslices. */
> > -	sseu->max_slices = IS_GEN9_LP(dev_priv) ? 1 : 3;
> > -	sseu->max_subslices = IS_GEN9_LP(dev_priv) ? 3 : 4;
> > +	sseu->max_slices = IS_GEN9_LP(i915) ? 1 : 3;
> > +	sseu->max_subslices = IS_GEN9_LP(i915) ? 3 : 4;
> >   	sseu->max_eus_per_subslice = 8;
> >   	/*
> > @@ -442,12 +442,12 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	 * pair per subslice.
> >   	*/
> >   	sseu->has_slice_pg =
> > -		!IS_GEN9_LP(dev_priv) && hweight8(sseu->slice_mask) > 1;
> > +		!IS_GEN9_LP(i915) && hweight8(sseu->slice_mask) > 1;
> >   	sseu->has_subslice_pg =
> > -		IS_GEN9_LP(dev_priv) && intel_sseu_subslice_total(sseu) > 1;
> > +		IS_GEN9_LP(i915) && intel_sseu_subslice_total(sseu) > 1;
> >   	sseu->has_eu_pg = sseu->eu_per_subslice > 2;
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   #define IS_SS_DISABLED(ss)	(!(sseu->subslice_mask[0] & BIT(ss)))
> >   		info->has_pooled_eu = hweight8(sseu->subslice_mask[0]) == 3;
> > @@ -464,9 +464,9 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void broadwell_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	int s, ss;
> >   	u32 fuse2, subslice_mask, eu_disable[3]; /* s_max */
> > @@ -547,9 +547,9 @@ static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->has_eu_pg = 0;
> >   }
> > -static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> > +static void haswell_sseu_info_init(struct drm_i915_private *i915)
> >   {
> > -	struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu;
> > +	struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu;
> >   	u32 fuse1;
> >   	int s, ss;
> > @@ -557,9 +557,9 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	 * There isn't a register to tell us how many slices/subslices. We
> >   	 * work off the PCI-ids here.
> >   	 */
> > -	switch (INTEL_INFO(dev_priv)->gt) {
> > +	switch (INTEL_INFO(i915)->gt) {
> >   	default:
> > -		MISSING_CASE(INTEL_INFO(dev_priv)->gt);
> > +		MISSING_CASE(INTEL_INFO(i915)->gt);
> >   		/* fall through */
> >   	case 1:
> >   		sseu->slice_mask = BIT(0);
> > @@ -612,7 +612,7 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv)
> >   	sseu->has_eu_pg = 0;
> >   }
> > -static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv)
> > +static u32 read_reference_ts_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 ts_override = I915_READ(GEN9_TIMESTAMP_OVERRIDE);
> >   	u32 base_freq, frac_freq;
> > @@ -629,7 +629,7 @@ static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv)
> >   	return base_freq + frac_freq;
> >   }
> > -static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> > +static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *i915,
> >   					u32 rpm_config_reg)
> >   {
> >   	u32 f19_2_mhz = 19200;
> > @@ -649,7 +649,7 @@ static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> > +static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *i915,
> >   					u32 rpm_config_reg)
> >   {
> >   	u32 f19_2_mhz = 19200;
> > @@ -675,21 +675,21 @@ static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> > +static u32 read_timestamp_frequency(struct drm_i915_private *i915)
> >   {
> >   	u32 f12_5_mhz = 12500;
> >   	u32 f19_2_mhz = 19200;
> >   	u32 f24_mhz = 24000;
> > -	if (INTEL_GEN(dev_priv) <= 4) {
> > +	if (INTEL_GEN(i915) <= 4) {
> >   		/* PRMs say:
> >   		 *
> >   		 *     "The value in this register increments once every 16
> >   		 *      hclks." (through the “Clocking Configuration”
> >   		 *      (“CLKCFG”) MCHBAR register)
> >   		 */
> > -		return dev_priv->rawclk_freq / 16;
> > -	} else if (INTEL_GEN(dev_priv) <= 8) {
> > +		return i915->rawclk_freq / 16;
> > +	} else if (INTEL_GEN(i915) <= 8) {
> >   		/* PRMs say:
> >   		 *
> >   		 *     "The PCU TSC counts 10ns increments; this timestamp
> > @@ -697,14 +697,14 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> >   		 *      rolling over every 1.5 hours).
> >   		 */
> >   		return f12_5_mhz;
> > -	} else if (INTEL_GEN(dev_priv) <= 9) {
> > +	} else if (INTEL_GEN(i915) <= 9) {
> >   		u32 ctc_reg = I915_READ(CTC_MODE);
> >   		u32 freq = 0;
> >   		if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) {
> > -			freq = read_reference_ts_freq(dev_priv);
> > +			freq = read_reference_ts_freq(i915);
> >   		} else {
> > -			freq = IS_GEN9_LP(dev_priv) ? f19_2_mhz : f24_mhz;
> > +			freq = IS_GEN9_LP(i915) ? f19_2_mhz : f24_mhz;
> >   			/* Now figure out how the command stream's timestamp
> >   			 * register increments from this frequency (it might
> > @@ -715,7 +715,7 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> >   		}
> >   		return freq;
> > -	} else if (INTEL_GEN(dev_priv) <= 11) {
> > +	} else if (INTEL_GEN(i915) <= 11) {
> >   		u32 ctc_reg = I915_READ(CTC_MODE);
> >   		u32 freq = 0;
> > @@ -725,15 +725,15 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv)
> >   		 * tells us which one we should use.
> >   		 */
> >   		if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) {
> > -			freq = read_reference_ts_freq(dev_priv);
> > +			freq = read_reference_ts_freq(i915);
> >   		} else {
> >   			u32 rpm_config_reg = I915_READ(RPM_CONFIG0);
> > -			if (INTEL_GEN(dev_priv) <= 10)
> > -				freq = gen10_get_crystal_clock_freq(dev_priv,
> > +			if (INTEL_GEN(i915) <= 10)
> > +				freq = gen10_get_crystal_clock_freq(i915,
> >   								rpm_config_reg);
> >   			else
> > -				freq = gen11_get_crystal_clock_freq(dev_priv,
> > +				freq = gen11_get_crystal_clock_freq(i915,
> >   								rpm_config_reg);
> >   			/* Now figure out how the command stream's timestamp
> > @@ -841,7 +841,7 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915)
> >   /**
> >    * intel_device_info_runtime_init - initialize runtime info
> > - * @dev_priv: the i915 device
> > + * @i915: the i915 device
> >    *
> >    * Determine various intel_device_info fields at runtime.
> >    *
> > @@ -855,16 +855,16 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915)
> >    *   - after the PCH has been detected,
> >    *   - before the first usage of the fields it can tweak.
> >    */
> > -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> > +void intel_device_info_runtime_init(struct drm_i915_private *i915)
> >   {
> > -	struct intel_device_info *info = mkwrite_device_info(dev_priv);
> > -	struct intel_runtime_info *runtime = RUNTIME_INFO(dev_priv);
> > +	struct intel_device_info *info = mkwrite_device_info(i915);
> > +	struct intel_runtime_info *runtime = RUNTIME_INFO(i915);
> >   	enum pipe pipe;
> > -	if (INTEL_GEN(dev_priv) >= 10) {
> > -		for_each_pipe(dev_priv, pipe)
> > +	if (INTEL_GEN(i915) >= 10) {
> > +		for_each_pipe(i915, pipe)
> >   			runtime->num_scalers[pipe] = 2;
> > -	} else if (IS_GEN(dev_priv, 9)) {
> > +	} else if (IS_GEN(i915, 9)) {
> >   		runtime->num_scalers[PIPE_A] = 2;
> >   		runtime->num_scalers[PIPE_B] = 2;
> >   		runtime->num_scalers[PIPE_C] = 1;
> > @@ -872,13 +872,13 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> >   	BUILD_BUG_ON(BITS_PER_TYPE(intel_engine_mask_t) < I915_NUM_ENGINES);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		for_each_pipe(dev_priv, pipe)
> > +	if (INTEL_GEN(i915) >= 11)
> > +		for_each_pipe(i915, pipe)
> >   			runtime->num_sprites[pipe] = 6;
> > -	else if (IS_GEN(dev_priv, 10) || IS_GEMINILAKE(dev_priv))
> > -		for_each_pipe(dev_priv, pipe)
> > +	else if (IS_GEN(i915, 10) || IS_GEMINILAKE(i915))
> > +		for_each_pipe(i915, pipe)
> >   			runtime->num_sprites[pipe] = 3;
> > -	else if (IS_BROXTON(dev_priv)) {
> > +	else if (IS_BROXTON(i915)) {
> >   		/*
> >   		 * Skylake and Broxton currently don't expose the topmost plane as its
> >   		 * use is exclusive with the legacy cursor and we only want to expose
> > @@ -891,20 +891,20 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> >   		runtime->num_sprites[PIPE_A] = 2;
> >   		runtime->num_sprites[PIPE_B] = 2;
> >   		runtime->num_sprites[PIPE_C] = 1;
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		for_each_pipe(dev_priv, pipe)
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		for_each_pipe(i915, pipe)
> >   			runtime->num_sprites[pipe] = 2;
> > -	} else if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) {
> > -		for_each_pipe(dev_priv, pipe)
> > +	} else if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) {
> > +		for_each_pipe(i915, pipe)
> >   			runtime->num_sprites[pipe] = 1;
> >   	}
> >   	if (i915_modparams.disable_display) {
> >   		DRM_INFO("Display disabled (module parameter)\n");
> >   		info->num_pipes = 0;
> > -	} else if (HAS_DISPLAY(dev_priv) &&
> > -		   (IS_GEN_RANGE(dev_priv, 7, 8)) &&
> > -		   HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_DISPLAY(i915) &&
> > +		   (IS_GEN_RANGE(i915, 7, 8)) &&
> > +		   HAS_PCH_SPLIT(i915)) {
> >   		u32 fuse_strap = I915_READ(FUSE_STRAP);
> >   		u32 sfuse_strap = I915_READ(SFUSE_STRAP);
> > @@ -919,7 +919,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> >   		 */
> >   		if (fuse_strap & ILK_INTERNAL_DISPLAY_DISABLE ||
> >   		    sfuse_strap & SFUSE_STRAP_DISPLAY_DISABLED ||
> > -		    (HAS_PCH_CPT(dev_priv) &&
> > +		    (HAS_PCH_CPT(i915) &&
> >   		     !(sfuse_strap & SFUSE_STRAP_FUSE_LOCK))) {
> >   			DRM_INFO("Display fused off, disabling\n");
> >   			info->num_pipes = 0;
> > @@ -927,7 +927,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> >   			DRM_INFO("PipeC fused off\n");
> >   			info->num_pipes -= 1;
> >   		}
> > -	} else if (HAS_DISPLAY(dev_priv) && INTEL_GEN(dev_priv) >= 9) {
> > +	} else if (HAS_DISPLAY(i915) && INTEL_GEN(i915) >= 9) {
> >   		u32 dfsm = I915_READ(SKL_DFSM);
> >   		u8 disabled_mask = 0;
> >   		bool invalid;
> > @@ -961,26 +961,26 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv)
> >   	}
> >   	/* Initialize slice/subslice/EU info */
> > -	if (IS_HASWELL(dev_priv))
> > -		haswell_sseu_info_init(dev_priv);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		cherryview_sseu_info_init(dev_priv);
> > -	else if (IS_BROADWELL(dev_priv))
> > -		broadwell_sseu_info_init(dev_priv);
> > -	else if (IS_GEN(dev_priv, 9))
> > -		gen9_sseu_info_init(dev_priv);
> > -	else if (IS_GEN(dev_priv, 10))
> > -		gen10_sseu_info_init(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 11)
> > -		gen11_sseu_info_init(dev_priv);
> > -
> > -	if (IS_GEN(dev_priv, 6) && intel_vtd_active()) {
> > +	if (IS_HASWELL(i915))
> > +		haswell_sseu_info_init(i915);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		cherryview_sseu_info_init(i915);
> > +	else if (IS_BROADWELL(i915))
> > +		broadwell_sseu_info_init(i915);
> > +	else if (IS_GEN(i915, 9))
> > +		gen9_sseu_info_init(i915);
> > +	else if (IS_GEN(i915, 10))
> > +		gen10_sseu_info_init(i915);
> > +	else if (INTEL_GEN(i915) >= 11)
> > +		gen11_sseu_info_init(i915);
> > +
> > +	if (IS_GEN(i915, 6) && intel_vtd_active()) {
> >   		DRM_INFO("Disabling ppGTT for VT-d support\n");
> >   		info->ppgtt_type = INTEL_PPGTT_NONE;
> >   	}
> >   	/* Initialize command stream timestamp frequency */
> > -	runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(dev_priv);
> > +	runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(i915);
> >   }
> >   void intel_driver_caps_print(const struct intel_driver_caps *caps,
> > @@ -997,16 +997,16 @@ void intel_driver_caps_print(const struct intel_driver_caps *caps,
> >    * this point (but later we need to prune the forcewake domains for engines that
> >    * are indeed fused off).
> >    */
> > -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> > +void intel_device_info_init_mmio(struct drm_i915_private *i915)
> >   {
> > -	struct intel_device_info *info = mkwrite_device_info(dev_priv);
> > +	struct intel_device_info *info = mkwrite_device_info(i915);
> >   	unsigned int logical_vdbox = 0;
> >   	unsigned int i;
> >   	u32 media_fuse;
> >   	u16 vdbox_mask;
> >   	u16 vebox_mask;
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return;
> >   	media_fuse = ~I915_READ(GEN11_GT_VEBOX_VDBOX_DISABLE);
> > @@ -1016,7 +1016,7 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> >   		      GEN11_GT_VEBOX_DISABLE_SHIFT;
> >   	for (i = 0; i < I915_MAX_VCS; i++) {
> > -		if (!HAS_ENGINE(dev_priv, _VCS(i)))
> > +		if (!HAS_ENGINE(i915, _VCS(i)))
> >   			continue;
> >   		if (!(BIT(i) & vdbox_mask)) {
> > @@ -1030,14 +1030,14 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> >   		 * hooked up to an SFC (Scaler & Format Converter) unit.
> >   		 */
> >   		if (logical_vdbox++ % 2 == 0)
> > -			RUNTIME_INFO(dev_priv)->vdbox_sfc_access |= BIT(i);
> > +			RUNTIME_INFO(i915)->vdbox_sfc_access |= BIT(i);
> >   	}
> >   	DRM_DEBUG_DRIVER("vdbox enable: %04x, instances: %04lx\n",
> > -			 vdbox_mask, VDBOX_MASK(dev_priv));
> > -	GEM_BUG_ON(vdbox_mask != VDBOX_MASK(dev_priv));
> > +			 vdbox_mask, VDBOX_MASK(i915));
> > +	GEM_BUG_ON(vdbox_mask != VDBOX_MASK(i915));
> >   	for (i = 0; i < I915_MAX_VECS; i++) {
> > -		if (!HAS_ENGINE(dev_priv, _VECS(i)))
> > +		if (!HAS_ENGINE(i915, _VECS(i)))
> >   			continue;
> >   		if (!(BIT(i) & vebox_mask)) {
> > @@ -1046,6 +1046,6 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv)
> >   		}
> >   	}
> >   	DRM_DEBUG_DRIVER("vebox enable: %04x, instances: %04lx\n",
> > -			 vebox_mask, VEBOX_MASK(dev_priv));
> > -	GEM_BUG_ON(vebox_mask != VEBOX_MASK(dev_priv));
> > +			 vebox_mask, VEBOX_MASK(i915));
> > +	GEM_BUG_ON(vebox_mask != VEBOX_MASK(i915));
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_device_info.h b/drivers/gpu/drm/i915/intel_device_info.h
> > index 3ea953a230b3..93e5cc95e5c0 100644
> > --- a/drivers/gpu/drm/i915/intel_device_info.h
> > +++ b/drivers/gpu/drm/i915/intel_device_info.h
> > @@ -218,8 +218,8 @@ struct intel_driver_caps {
> >   const char *intel_platform_name(enum intel_platform platform);
> > -void intel_device_info_subplatform_init(struct drm_i915_private *dev_priv);
> > -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv);
> > +void intel_device_info_subplatform_init(struct drm_i915_private *i915);
> > +void intel_device_info_runtime_init(struct drm_i915_private *i915);
> >   void intel_device_info_dump_flags(const struct intel_device_info *info,
> >   				  struct drm_printer *p);
> >   void intel_device_info_dump_runtime(const struct intel_runtime_info *info,
> > @@ -227,7 +227,7 @@ void intel_device_info_dump_runtime(const struct intel_runtime_info *info,
> >   void intel_device_info_dump_topology(const struct sseu_dev_info *sseu,
> >   				     struct drm_printer *p);
> > -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv);
> > +void intel_device_info_init_mmio(struct drm_i915_private *i915);
> >   void intel_driver_caps_print(const struct intel_driver_caps *caps,
> >   			     struct drm_printer *p);
> > diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
> > index 1b1ddb48ca7a..be26ad9fc194 100644
> > --- a/drivers/gpu/drm/i915/intel_display.c
> > +++ b/drivers/gpu/drm/i915/intel_display.c
> > @@ -156,24 +156,24 @@ struct intel_limit {
> >   };
> >   /* returns HPLL frequency in kHz */
> > -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv)
> > +int vlv_get_hpll_vco(struct drm_i915_private *i915)
> >   {
> >   	int hpll_freq, vco_freq[] = { 800, 1600, 2000, 2400 };
> >   	/* Obtain SKU information */
> > -	hpll_freq = vlv_cck_read(dev_priv, CCK_FUSE_REG) &
> > +	hpll_freq = vlv_cck_read(i915, CCK_FUSE_REG) &
> >   		CCK_FUSE_HPLL_FREQ_MASK;
> >   	return vco_freq[hpll_freq] * 1000;
> >   }
> > -int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> > +int vlv_get_cck_clock(struct drm_i915_private *i915,
> >   		      const char *name, u32 reg, int ref_freq)
> >   {
> >   	u32 val;
> >   	int divider;
> > -	val = vlv_cck_read(dev_priv, reg);
> > +	val = vlv_cck_read(i915, reg);
> >   	divider = val & CCK_FREQUENCY_VALUES;
> >   	WARN((val & CCK_FREQUENCY_STATUS) !=
> > @@ -183,42 +183,42 @@ int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> >   	return DIV_ROUND_CLOSEST(ref_freq << 1, divider + 1);
> >   }
> > -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv,
> > +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915,
> >   			   const char *name, u32 reg)
> >   {
> >   	int hpll;
> > -	vlv_cck_get(dev_priv);
> > +	vlv_cck_get(i915);
> > -	if (dev_priv->hpll_freq == 0)
> > -		dev_priv->hpll_freq = vlv_get_hpll_vco(dev_priv);
> > +	if (i915->hpll_freq == 0)
> > +		i915->hpll_freq = vlv_get_hpll_vco(i915);
> > -	hpll = vlv_get_cck_clock(dev_priv, name, reg, dev_priv->hpll_freq);
> > +	hpll = vlv_get_cck_clock(i915, name, reg, i915->hpll_freq);
> > -	vlv_cck_put(dev_priv);
> > +	vlv_cck_put(i915);
> >   	return hpll;
> >   }
> > -static void intel_update_czclk(struct drm_i915_private *dev_priv)
> > +static void intel_update_czclk(struct drm_i915_private *i915)
> >   {
> > -	if (!(IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)))
> > +	if (!(IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)))
> >   		return;
> > -	dev_priv->czclk_freq = vlv_get_cck_clock_hpll(dev_priv, "czclk",
> > +	i915->czclk_freq = vlv_get_cck_clock_hpll(i915, "czclk",
> >   						      CCK_CZ_CLOCK_CONTROL);
> > -	DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", dev_priv->czclk_freq);
> > +	DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", i915->czclk_freq);
> >   }
> >   static inline u32 /* units of 100MHz */
> > -intel_fdi_link_freq(struct drm_i915_private *dev_priv,
> > +intel_fdi_link_freq(struct drm_i915_private *i915,
> >   		    const struct intel_crtc_state *pipe_config)
> >   {
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		return pipe_config->port_clock; /* SPLL */
> >   	else
> > -		return dev_priv->fdi_pll_freq;
> > +		return i915->fdi_pll_freq;
> >   }
> >   static const struct intel_limit intel_limits_i8xx_dac = {
> > @@ -488,7 +488,7 @@ static const struct intel_limit intel_limits_bxt = {
> >   /* WA Display #0827: Gen9:all */
> >   static void
> > -skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable)
> > +skl_wa_827(struct drm_i915_private *i915, int pipe, bool enable)
> >   {
> >   	if (enable)
> >   		I915_WRITE(CLKGATE_DIS_PSL(pipe),
> > @@ -502,7 +502,7 @@ skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable)
> >   /* Wa_2006604312:icl */
> >   static void
> > -icl_wa_scalerclkgating(struct drm_i915_private *dev_priv, enum pipe pipe,
> > +icl_wa_scalerclkgating(struct drm_i915_private *i915, enum pipe pipe,
> >   		       bool enable)
> >   {
> >   	if (enable)
> > @@ -588,7 +588,7 @@ int chv_calc_dpll_params(int refclk, struct dpll *clock)
> >    * Returns whether the given set of divisors are valid for a given refclk with
> >    * the given connectors.
> >    */
> > -static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv,
> > +static bool intel_PLL_is_valid(struct drm_i915_private *i915,
> >   			       const struct intel_limit *limit,
> >   			       const struct dpll *clock)
> >   {
> > @@ -601,13 +601,13 @@ static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv,
> >   	if (clock->m1  < limit->m1.min  || limit->m1.max  < clock->m1)
> >   		INTELPllInvalid("m1 out of range\n");
> > -	if (!IS_PINEVIEW(dev_priv) && !IS_VALLEYVIEW(dev_priv) &&
> > -	    !IS_CHERRYVIEW(dev_priv) && !IS_GEN9_LP(dev_priv))
> > +	if (!IS_PINEVIEW(i915) && !IS_VALLEYVIEW(i915) &&
> > +	    !IS_CHERRYVIEW(i915) && !IS_GEN9_LP(i915))
> >   		if (clock->m1 <= clock->m2)
> >   			INTELPllInvalid("m1 <= m2\n");
> > -	if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> > -	    !IS_GEN9_LP(dev_priv)) {
> > +	if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> > +	    !IS_GEN9_LP(i915)) {
> >   		if (clock->p < limit->p.min || limit->p.max < clock->p)
> >   			INTELPllInvalid("p out of range\n");
> >   		if (clock->m < limit->m.min || limit->m.max < clock->m)
> > @@ -630,7 +630,7 @@ i9xx_select_p2_div(const struct intel_limit *limit,
> >   		   const struct intel_crtc_state *crtc_state,
> >   		   int target)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> >   		/*
> > @@ -638,7 +638,7 @@ i9xx_select_p2_div(const struct intel_limit *limit,
> >   		 * We haven't figured out how to reliably set up different
> >   		 * single/dual channel state, if we even can.
> >   		 */
> > -		if (intel_is_dual_link_lvds(dev_priv))
> > +		if (intel_is_dual_link_lvds(i915))
> >   			return limit->p2.p2_fast;
> >   		else
> >   			return limit->p2.p2_slow;
> > @@ -1016,22 +1016,22 @@ bool intel_crtc_active(struct intel_crtc *crtc)
> >   		crtc->config->base.adjusted_mode.crtc_clock;
> >   }
> > -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv,
> > +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915,
> >   					     enum pipe pipe)
> >   {
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	return crtc->config->cpu_transcoder;
> >   }
> > -static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv,
> > +static bool pipe_scanline_is_moving(struct drm_i915_private *i915,
> >   				    enum pipe pipe)
> >   {
> >   	i915_reg_t reg = PIPEDSL(pipe);
> >   	u32 line1, line2;
> >   	u32 line_mask;
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		line_mask = DSL_LINEMASK_GEN2;
> >   	else
> >   		line_mask = DSL_LINEMASK_GEN3;
> > @@ -1045,11 +1045,11 @@ static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv,
> >   static void wait_for_pipe_scanline_moving(struct intel_crtc *crtc, bool state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* Wait for the display line to settle/start moving */
> > -	if (wait_for(pipe_scanline_is_moving(dev_priv, pipe) == state, 100))
> > +	if (wait_for(pipe_scanline_is_moving(i915, pipe) == state, 100))
> >   		DRM_ERROR("pipe %c scanline %s wait timed out\n",
> >   			  pipe_name(pipe), onoff(state));
> >   }
> > @@ -1068,14 +1068,14 @@ static void
> >   intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> >   		i915_reg_t reg = PIPECONF(cpu_transcoder);
> >   		/* Wait for the Pipe State to go off */
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    reg, I965_PIPECONF_ACTIVE, 0,
> >   					    100))
> >   			WARN(1, "pipe_off wait timed out\n");
> > @@ -1085,7 +1085,7 @@ intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state)
> >   }
> >   /* Only for pre-ILK configs */
> > -void assert_pll(struct drm_i915_private *dev_priv,
> > +void assert_pll(struct drm_i915_private *i915,
> >   		enum pipe pipe, bool state)
> >   {
> >   	u32 val;
> > @@ -1099,14 +1099,14 @@ void assert_pll(struct drm_i915_private *dev_priv,
> >   }
> >   /* XXX: the dsi pll is shared between MIPI DSI ports */
> > -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state)
> > +void assert_dsi_pll(struct drm_i915_private *i915, bool state)
> >   {
> >   	u32 val;
> >   	bool cur_state;
> > -	vlv_cck_get(dev_priv);
> > -	val = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> > -	vlv_cck_put(dev_priv);
> > +	vlv_cck_get(i915);
> > +	val = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> > +	vlv_cck_put(i915);
> >   	cur_state = val & DSI_PLL_VCO_EN;
> >   	I915_STATE_WARN(cur_state != state,
> > @@ -1114,14 +1114,14 @@ void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state)
> >   			onoff(state), onoff(cur_state));
> >   }
> > -static void assert_fdi_tx(struct drm_i915_private *dev_priv,
> > +static void assert_fdi_tx(struct drm_i915_private *i915,
> >   			  enum pipe pipe, bool state)
> >   {
> >   	bool cur_state;
> > -	enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv,
> > +	enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915,
> >   								      pipe);
> > -	if (HAS_DDI(dev_priv)) {
> > +	if (HAS_DDI(i915)) {
> >   		/* DDI does not have a specific FDI_TX register */
> >   		u32 val = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder));
> >   		cur_state = !!(val & TRANS_DDI_FUNC_ENABLE);
> > @@ -1136,7 +1136,7 @@ static void assert_fdi_tx(struct drm_i915_private *dev_priv,
> >   #define assert_fdi_tx_enabled(d, p) assert_fdi_tx(d, p, true)
> >   #define assert_fdi_tx_disabled(d, p) assert_fdi_tx(d, p, false)
> > -static void assert_fdi_rx(struct drm_i915_private *dev_priv,
> > +static void assert_fdi_rx(struct drm_i915_private *i915,
> >   			  enum pipe pipe, bool state)
> >   {
> >   	u32 val;
> > @@ -1151,24 +1151,24 @@ static void assert_fdi_rx(struct drm_i915_private *dev_priv,
> >   #define assert_fdi_rx_enabled(d, p) assert_fdi_rx(d, p, true)
> >   #define assert_fdi_rx_disabled(d, p) assert_fdi_rx(d, p, false)
> > -static void assert_fdi_tx_pll_enabled(struct drm_i915_private *dev_priv,
> > +static void assert_fdi_tx_pll_enabled(struct drm_i915_private *i915,
> >   				      enum pipe pipe)
> >   {
> >   	u32 val;
> >   	/* ILK FDI PLL is always enabled */
> > -	if (IS_GEN(dev_priv, 5))
> > +	if (IS_GEN(i915, 5))
> >   		return;
> >   	/* On Haswell, DDI ports are responsible for the FDI PLL setup */
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		return;
> >   	val = I915_READ(FDI_TX_CTL(pipe));
> >   	I915_STATE_WARN(!(val & FDI_TX_PLL_ENABLE), "FDI TX PLL assertion failure, should be active but is disabled\n");
> >   }
> > -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> > +void assert_fdi_rx_pll(struct drm_i915_private *i915,
> >   		       enum pipe pipe, bool state)
> >   {
> >   	u32 val;
> > @@ -1181,17 +1181,17 @@ void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> >   			onoff(state), onoff(cur_state));
> >   }
> > -void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +void assert_panel_unlocked(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	i915_reg_t pp_reg;
> >   	u32 val;
> >   	enum pipe panel_pipe = INVALID_PIPE;
> >   	bool locked = true;
> > -	if (WARN_ON(HAS_DDI(dev_priv)))
> > +	if (WARN_ON(HAS_DDI(i915)))
> >   		return;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		u32 port_sel;
> >   		pp_reg = PP_CONTROL(0);
> > @@ -1199,22 +1199,22 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   		switch (port_sel) {
> >   		case PANEL_PORT_SELECT_LVDS:
> > -			intel_lvds_port_enabled(dev_priv, PCH_LVDS, &panel_pipe);
> > +			intel_lvds_port_enabled(i915, PCH_LVDS, &panel_pipe);
> >   			break;
> >   		case PANEL_PORT_SELECT_DPA:
> > -			intel_dp_port_enabled(dev_priv, DP_A, PORT_A, &panel_pipe);
> > +			intel_dp_port_enabled(i915, DP_A, PORT_A, &panel_pipe);
> >   			break;
> >   		case PANEL_PORT_SELECT_DPC:
> > -			intel_dp_port_enabled(dev_priv, PCH_DP_C, PORT_C, &panel_pipe);
> > +			intel_dp_port_enabled(i915, PCH_DP_C, PORT_C, &panel_pipe);
> >   			break;
> >   		case PANEL_PORT_SELECT_DPD:
> > -			intel_dp_port_enabled(dev_priv, PCH_DP_D, PORT_D, &panel_pipe);
> > +			intel_dp_port_enabled(i915, PCH_DP_D, PORT_D, &panel_pipe);
> >   			break;
> >   		default:
> >   			MISSING_CASE(port_sel);
> >   			break;
> >   		}
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		/* presumably write lock depends on pipe, not port select */
> >   		pp_reg = PP_CONTROL(pipe);
> >   		panel_pipe = pipe;
> > @@ -1225,7 +1225,7 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   		port_sel = I915_READ(PP_ON_DELAYS(0)) & PANEL_PORT_SELECT_MASK;
> >   		WARN_ON(port_sel != PANEL_PORT_SELECT_LVDS);
> > -		intel_lvds_port_enabled(dev_priv, LVDS, &panel_pipe);
> > +		intel_lvds_port_enabled(i915, LVDS, &panel_pipe);
> >   	}
> >   	val = I915_READ(pp_reg);
> > @@ -1238,26 +1238,26 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	     pipe_name(pipe));
> >   }
> > -void assert_pipe(struct drm_i915_private *dev_priv,
> > +void assert_pipe(struct drm_i915_private *i915,
> >   		 enum pipe pipe, bool state)
> >   {
> >   	bool cur_state;
> > -	enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv,
> > +	enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915,
> >   								      pipe);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	/* we keep both pipes enabled on 830 */
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		state = true;
> >   	power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (wakeref) {
> >   		u32 val = I915_READ(PIPECONF(cpu_transcoder));
> >   		cur_state = !!(val & PIPECONF_ENABLE);
> > -		intel_display_power_put(dev_priv, power_domain, wakeref);
> > +		intel_display_power_put(i915, power_domain, wakeref);
> >   	} else {
> >   		cur_state = false;
> >   	}
> > @@ -1284,10 +1284,10 @@ static void assert_plane(struct intel_plane *plane, bool state)
> >   static void assert_planes_disabled(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_plane *plane;
> > -	for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane)
> > +	for_each_intel_plane_on_crtc(&i915->drm, crtc, plane)
> >   		assert_plane_disabled(plane);
> >   }
> > @@ -1297,7 +1297,7 @@ static void assert_vblank_disabled(struct drm_crtc *crtc)
> >   		drm_crtc_vblank_put(crtc);
> >   }
> > -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> > +void assert_pch_transcoder_disabled(struct drm_i915_private *i915,
> >   				    enum pipe pipe)
> >   {
> >   	u32 val;
> > @@ -1310,78 +1310,78 @@ void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> >   	     pipe_name(pipe));
> >   }
> > -static void assert_pch_dp_disabled(struct drm_i915_private *dev_priv,
> > +static void assert_pch_dp_disabled(struct drm_i915_private *i915,
> >   				   enum pipe pipe, enum port port,
> >   				   i915_reg_t dp_reg)
> >   {
> >   	enum pipe port_pipe;
> >   	bool state;
> > -	state = intel_dp_port_enabled(dev_priv, dp_reg, port, &port_pipe);
> > +	state = intel_dp_port_enabled(i915, dp_reg, port, &port_pipe);
> >   	I915_STATE_WARN(state && port_pipe == pipe,
> >   			"PCH DP %c enabled on transcoder %c, should be disabled\n",
> >   			port_name(port), pipe_name(pipe));
> > -	I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B,
> > +	I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B,
> >   			"IBX PCH DP %c still using transcoder B\n",
> >   			port_name(port));
> >   }
> > -static void assert_pch_hdmi_disabled(struct drm_i915_private *dev_priv,
> > +static void assert_pch_hdmi_disabled(struct drm_i915_private *i915,
> >   				     enum pipe pipe, enum port port,
> >   				     i915_reg_t hdmi_reg)
> >   {
> >   	enum pipe port_pipe;
> >   	bool state;
> > -	state = intel_sdvo_port_enabled(dev_priv, hdmi_reg, &port_pipe);
> > +	state = intel_sdvo_port_enabled(i915, hdmi_reg, &port_pipe);
> >   	I915_STATE_WARN(state && port_pipe == pipe,
> >   			"PCH HDMI %c enabled on transcoder %c, should be disabled\n",
> >   			port_name(port), pipe_name(pipe));
> > -	I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B,
> > +	I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B,
> >   			"IBX PCH HDMI %c still using transcoder B\n",
> >   			port_name(port));
> >   }
> > -static void assert_pch_ports_disabled(struct drm_i915_private *dev_priv,
> > +static void assert_pch_ports_disabled(struct drm_i915_private *i915,
> >   				      enum pipe pipe)
> >   {
> >   	enum pipe port_pipe;
> > -	assert_pch_dp_disabled(dev_priv, pipe, PORT_B, PCH_DP_B);
> > -	assert_pch_dp_disabled(dev_priv, pipe, PORT_C, PCH_DP_C);
> > -	assert_pch_dp_disabled(dev_priv, pipe, PORT_D, PCH_DP_D);
> > +	assert_pch_dp_disabled(i915, pipe, PORT_B, PCH_DP_B);
> > +	assert_pch_dp_disabled(i915, pipe, PORT_C, PCH_DP_C);
> > +	assert_pch_dp_disabled(i915, pipe, PORT_D, PCH_DP_D);
> > -	I915_STATE_WARN(intel_crt_port_enabled(dev_priv, PCH_ADPA, &port_pipe) &&
> > +	I915_STATE_WARN(intel_crt_port_enabled(i915, PCH_ADPA, &port_pipe) &&
> >   			port_pipe == pipe,
> >   			"PCH VGA enabled on transcoder %c, should be disabled\n",
> >   			pipe_name(pipe));
> > -	I915_STATE_WARN(intel_lvds_port_enabled(dev_priv, PCH_LVDS, &port_pipe) &&
> > +	I915_STATE_WARN(intel_lvds_port_enabled(i915, PCH_LVDS, &port_pipe) &&
> >   			port_pipe == pipe,
> >   			"PCH LVDS enabled on transcoder %c, should be disabled\n",
> >   			pipe_name(pipe));
> >   	/* PCH SDVOB multiplex with HDMIB */
> > -	assert_pch_hdmi_disabled(dev_priv, pipe, PORT_B, PCH_HDMIB);
> > -	assert_pch_hdmi_disabled(dev_priv, pipe, PORT_C, PCH_HDMIC);
> > -	assert_pch_hdmi_disabled(dev_priv, pipe, PORT_D, PCH_HDMID);
> > +	assert_pch_hdmi_disabled(i915, pipe, PORT_B, PCH_HDMIB);
> > +	assert_pch_hdmi_disabled(i915, pipe, PORT_C, PCH_HDMIC);
> > +	assert_pch_hdmi_disabled(i915, pipe, PORT_D, PCH_HDMID);
> >   }
> >   static void _vlv_enable_pll(struct intel_crtc *crtc,
> >   			    const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll);
> >   	POSTING_READ(DPLL(pipe));
> >   	udelay(150);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DPLL(pipe),
> >   				    DPLL_LOCK_VLV,
> >   				    DPLL_LOCK_VLV,
> > @@ -1392,13 +1392,13 @@ static void _vlv_enable_pll(struct intel_crtc *crtc,
> >   static void vlv_enable_pll(struct intel_crtc *crtc,
> >   			   const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> > -	assert_pipe_disabled(dev_priv, pipe);
> > +	assert_pipe_disabled(i915, pipe);
> >   	/* PLL is protected by panel, make sure we can write it */
> > -	assert_panel_unlocked(dev_priv, pipe);
> > +	assert_panel_unlocked(i915, pipe);
> >   	if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE)
> >   		_vlv_enable_pll(crtc, pipe_config);
> > @@ -1411,19 +1411,19 @@ static void vlv_enable_pll(struct intel_crtc *crtc,
> >   static void _chv_enable_pll(struct intel_crtc *crtc,
> >   			    const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum dpio_channel port = vlv_pipe_to_channel(pipe);
> >   	u32 tmp;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Enable back the 10bit clock to display controller */
> > -	tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port));
> > +	tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port));
> >   	tmp |= DPIO_DCLKP_EN;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), tmp);
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), tmp);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   	/*
> >   	 * Need to wait > 100ns between dclkp clock enable bit and PLL enable.
> > @@ -1434,7 +1434,7 @@ static void _chv_enable_pll(struct intel_crtc *crtc,
> >   	I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll);
> >   	/* Check PLL is locked */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DPLL(pipe), DPLL_LOCK_VLV, DPLL_LOCK_VLV,
> >   				    1))
> >   		DRM_ERROR("PLL %d failed to lock\n", pipe);
> > @@ -1443,13 +1443,13 @@ static void _chv_enable_pll(struct intel_crtc *crtc,
> >   static void chv_enable_pll(struct intel_crtc *crtc,
> >   			   const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> > -	assert_pipe_disabled(dev_priv, pipe);
> > +	assert_pipe_disabled(i915, pipe);
> >   	/* PLL is protected by panel, make sure we can write it */
> > -	assert_panel_unlocked(dev_priv, pipe);
> > +	assert_panel_unlocked(i915, pipe);
> >   	if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE)
> >   		_chv_enable_pll(crtc, pipe_config);
> > @@ -1464,7 +1464,7 @@ static void chv_enable_pll(struct intel_crtc *crtc,
> >   		I915_WRITE(CBR4_VLV, CBR_DPLLBMD_PIPE(pipe));
> >   		I915_WRITE(DPLL_MD(PIPE_B), pipe_config->dpll_hw_state.dpll_md);
> >   		I915_WRITE(CBR4_VLV, 0);
> > -		dev_priv->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md;
> > +		i915->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md;
> >   		/*
> >   		 * DPLLB VGA mode also seems to cause problems.
> > @@ -1477,27 +1477,27 @@ static void chv_enable_pll(struct intel_crtc *crtc,
> >   	}
> >   }
> > -static bool i9xx_has_pps(struct drm_i915_private *dev_priv)
> > +static bool i9xx_has_pps(struct drm_i915_private *i915)
> >   {
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		return false;
> > -	return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> > +	return IS_PINEVIEW(i915) || IS_MOBILE(i915);
> >   }
> >   static void i9xx_enable_pll(struct intel_crtc *crtc,
> >   			    const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	i915_reg_t reg = DPLL(crtc->pipe);
> >   	u32 dpll = crtc_state->dpll_hw_state.dpll;
> >   	int i;
> > -	assert_pipe_disabled(dev_priv, crtc->pipe);
> > +	assert_pipe_disabled(i915, crtc->pipe);
> >   	/* PLL is protected by panel, make sure we can write it */
> > -	if (i9xx_has_pps(dev_priv))
> > -		assert_panel_unlocked(dev_priv, crtc->pipe);
> > +	if (i9xx_has_pps(i915))
> > +		assert_panel_unlocked(i915, crtc->pipe);
> >   	/*
> >   	 * Apparently we need to have VGA mode enabled prior to changing
> > @@ -1511,7 +1511,7 @@ static void i9xx_enable_pll(struct intel_crtc *crtc,
> >   	POSTING_READ(reg);
> >   	udelay(150);
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		I915_WRITE(DPLL_MD(crtc->pipe),
> >   			   crtc_state->dpll_hw_state.dpll_md);
> >   	} else {
> > @@ -1534,26 +1534,26 @@ static void i9xx_enable_pll(struct intel_crtc *crtc,
> >   static void i9xx_disable_pll(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* Don't disable pipe or pipe PLLs if needed */
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		return;
> >   	/* Make sure the pipe isn't still relying on us */
> > -	assert_pipe_disabled(dev_priv, pipe);
> > +	assert_pipe_disabled(i915, pipe);
> >   	I915_WRITE(DPLL(pipe), DPLL_VGA_MODE_DIS);
> >   	POSTING_READ(DPLL(pipe));
> >   }
> > -static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +static void vlv_disable_pll(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	u32 val;
> >   	/* Make sure the pipe isn't still relying on us */
> > -	assert_pipe_disabled(dev_priv, pipe);
> > +	assert_pipe_disabled(i915, pipe);
> >   	val = DPLL_INTEGRATED_REF_CLK_VLV |
> >   		DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> > @@ -1564,13 +1564,13 @@ static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	POSTING_READ(DPLL(pipe));
> >   }
> > -static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +static void chv_disable_pll(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	enum dpio_channel port = vlv_pipe_to_channel(pipe);
> >   	u32 val;
> >   	/* Make sure the pipe isn't still relying on us */
> > -	assert_pipe_disabled(dev_priv, pipe);
> > +	assert_pipe_disabled(i915, pipe);
> >   	val = DPLL_SSC_REF_CLK_CHV |
> >   		DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> > @@ -1580,17 +1580,17 @@ static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	I915_WRITE(DPLL(pipe), val);
> >   	POSTING_READ(DPLL(pipe));
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Disable 10bit clock to display controller */
> > -	val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port));
> > +	val = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port));
> >   	val &= ~DPIO_DCLKP_EN;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), val);
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), val);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> > -void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> > +void vlv_wait_port_ready(struct drm_i915_private *i915,
> >   			 struct intel_digital_port *dport,
> >   			 unsigned int expected_mask)
> >   {
> > @@ -1615,7 +1615,7 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> >   		BUG();
> >   	}
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    dpll_reg, port_mask, expected_mask,
> >   				    1000))
> >   		WARN(1, "timed out waiting for port %c ready: got 0x%x, expected 0x%x\n",
> > @@ -1626,19 +1626,19 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> >   static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 val, pipeconf_val;
> >   	/* Make sure PCH DPLL is enabled */
> > -	assert_shared_dpll_enabled(dev_priv, crtc_state->shared_dpll);
> > +	assert_shared_dpll_enabled(i915, crtc_state->shared_dpll);
> >   	/* FDI must be feeding us bits for PCH ports */
> > -	assert_fdi_tx_enabled(dev_priv, pipe);
> > -	assert_fdi_rx_enabled(dev_priv, pipe);
> > +	assert_fdi_tx_enabled(i915, pipe);
> > +	assert_fdi_rx_enabled(i915, pipe);
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		/* Workaround: Set the timing override bit before enabling the
> >   		 * pch transcoder. */
> >   		reg = TRANS_CHICKEN2(pipe);
> > @@ -1651,7 +1651,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
> >   	val = I915_READ(reg);
> >   	pipeconf_val = I915_READ(PIPECONF(pipe));
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915)) {
> >   		/*
> >   		 * Make the BPC in transcoder be consistent with
> >   		 * that in pipeconf reg. For HDMI we must use 8bpc
> > @@ -1666,7 +1666,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
> >   	val &= ~TRANS_INTERLACE_MASK;
> >   	if ((pipeconf_val & PIPECONF_INTERLACE_MASK) == PIPECONF_INTERLACED_ILK) {
> > -		if (HAS_PCH_IBX(dev_priv) &&
> > +		if (HAS_PCH_IBX(i915) &&
> >   		    intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO))
> >   			val |= TRANS_LEGACY_INTERLACED_ILK;
> >   		else
> > @@ -1676,20 +1676,20 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s
> >   	}
> >   	I915_WRITE(reg, val | TRANS_ENABLE);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    reg, TRANS_STATE_ENABLE, TRANS_STATE_ENABLE,
> >   				    100))
> >   		DRM_ERROR("failed to enable transcoder %c\n", pipe_name(pipe));
> >   }
> > -static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> > +static void lpt_enable_pch_transcoder(struct drm_i915_private *i915,
> >   				      enum transcoder cpu_transcoder)
> >   {
> >   	u32 val, pipeconf_val;
> >   	/* FDI must be feeding us bits for PCH ports */
> > -	assert_fdi_tx_enabled(dev_priv, (enum pipe) cpu_transcoder);
> > -	assert_fdi_rx_enabled(dev_priv, PIPE_A);
> > +	assert_fdi_tx_enabled(i915, (enum pipe) cpu_transcoder);
> > +	assert_fdi_rx_enabled(i915, PIPE_A);
> >   	/* Workaround: set timing override bit. */
> >   	val = I915_READ(TRANS_CHICKEN2(PIPE_A));
> > @@ -1706,7 +1706,7 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> >   		val |= TRANS_PROGRESSIVE;
> >   	I915_WRITE(LPT_TRANSCONF, val);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    LPT_TRANSCONF,
> >   				    TRANS_STATE_ENABLE,
> >   				    TRANS_STATE_ENABLE,
> > @@ -1714,30 +1714,30 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv,
> >   		DRM_ERROR("Failed to enable PCH transcoder\n");
> >   }
> > -static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv,
> > +static void ironlake_disable_pch_transcoder(struct drm_i915_private *i915,
> >   					    enum pipe pipe)
> >   {
> >   	i915_reg_t reg;
> >   	u32 val;
> >   	/* FDI relies on the transcoder */
> > -	assert_fdi_tx_disabled(dev_priv, pipe);
> > -	assert_fdi_rx_disabled(dev_priv, pipe);
> > +	assert_fdi_tx_disabled(i915, pipe);
> > +	assert_fdi_rx_disabled(i915, pipe);
> >   	/* Ports must be off as well */
> > -	assert_pch_ports_disabled(dev_priv, pipe);
> > +	assert_pch_ports_disabled(i915, pipe);
> >   	reg = PCH_TRANSCONF(pipe);
> >   	val = I915_READ(reg);
> >   	val &= ~TRANS_ENABLE;
> >   	I915_WRITE(reg, val);
> >   	/* wait for PCH transcoder off, transcoder state */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    reg, TRANS_STATE_ENABLE, 0,
> >   				    50))
> >   		DRM_ERROR("failed to disable transcoder %c\n", pipe_name(pipe));
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		/* Workaround: Clear the timing override chicken bit again. */
> >   		reg = TRANS_CHICKEN2(pipe);
> >   		val = I915_READ(reg);
> > @@ -1746,7 +1746,7 @@ static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
> > +void lpt_disable_pch_transcoder(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -1754,7 +1754,7 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
> >   	val &= ~TRANS_ENABLE;
> >   	I915_WRITE(LPT_TRANSCONF, val);
> >   	/* wait for PCH transcoder off, transcoder state */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    LPT_TRANSCONF, TRANS_STATE_ENABLE, 0,
> >   				    50))
> >   		DRM_ERROR("Failed to disable PCH transcoder\n");
> > @@ -1767,9 +1767,9 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv)
> >   enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	if (HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_LPT(i915))
> >   		return PIPE_A;
> >   	else
> >   		return crtc->pipe;
> > @@ -1777,19 +1777,19 @@ enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc)
> >   static u32 intel_crtc_max_vblank_count(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	/*
> >   	 * On i965gm the hardware frame counter reads
> >   	 * zero when the TV encoder is enabled :(
> >   	 */
> > -	if (IS_I965GM(dev_priv) &&
> > +	if (IS_I965GM(i915) &&
> >   	    (crtc_state->output_types & BIT(INTEL_OUTPUT_TVOUT)))
> >   		return 0;
> > -	if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> >   		return 0xffffffff; /* full 32 bit counter */
> > -	else if (INTEL_GEN(dev_priv) >= 3)
> > +	else if (INTEL_GEN(i915) >= 3)
> >   		return 0xffffff; /* only 24 bits of frame count */
> >   	else
> >   		return 0; /* Gen2 doesn't have a hardware frame counter */
> > @@ -1807,7 +1807,7 @@ static void intel_crtc_vblank_on(const struct intel_crtc_state *crtc_state)
> >   static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = new_crtc_state->cpu_transcoder;
> >   	enum pipe pipe = crtc->pipe;
> >   	i915_reg_t reg;
> > @@ -1822,29 +1822,29 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> >   	 * a plane.  On ILK+ the pipe PLLs are integrated, so we don't
> >   	 * need the check.
> >   	 */
> > -	if (HAS_GMCH(dev_priv)) {
> > +	if (HAS_GMCH(i915)) {
> >   		if (intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI))
> > -			assert_dsi_pll_enabled(dev_priv);
> > +			assert_dsi_pll_enabled(i915);
> >   		else
> > -			assert_pll_enabled(dev_priv, pipe);
> > +			assert_pll_enabled(i915, pipe);
> >   	} else {
> >   		if (new_crtc_state->has_pch_encoder) {
> >   			/* if driving the PCH, we need FDI enabled */
> > -			assert_fdi_rx_pll_enabled(dev_priv,
> > +			assert_fdi_rx_pll_enabled(i915,
> >   						  intel_crtc_pch_transcoder(crtc));
> > -			assert_fdi_tx_pll_enabled(dev_priv,
> > +			assert_fdi_tx_pll_enabled(i915,
> >   						  (enum pipe) cpu_transcoder);
> >   		}
> >   		/* FIXME: assert CPU port conditions for SNB+ */
> >   	}
> > -	trace_intel_pipe_enable(dev_priv, pipe);
> > +	trace_intel_pipe_enable(i915, pipe);
> >   	reg = PIPECONF(cpu_transcoder);
> >   	val = I915_READ(reg);
> >   	if (val & PIPECONF_ENABLE) {
> >   		/* we keep both pipes enabled on 830 */
> > -		WARN_ON(!IS_I830(dev_priv));
> > +		WARN_ON(!IS_I830(i915));
> >   		return;
> >   	}
> > @@ -1865,7 +1865,7 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state)
> >   static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> >   	enum pipe pipe = crtc->pipe;
> >   	i915_reg_t reg;
> > @@ -1879,7 +1879,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> >   	 */
> >   	assert_planes_disabled(crtc);
> > -	trace_intel_pipe_disable(dev_priv, pipe);
> > +	trace_intel_pipe_disable(i915, pipe);
> >   	reg = PIPECONF(cpu_transcoder);
> >   	val = I915_READ(reg);
> > @@ -1894,7 +1894,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> >   		val &= ~PIPECONF_DOUBLE_WIDE;
> >   	/* Don't disable pipe or pipe PLLs if needed */
> > -	if (!IS_I830(dev_priv))
> > +	if (!IS_I830(i915))
> >   		val &= ~PIPECONF_ENABLE;
> >   	I915_WRITE(reg, val);
> > @@ -1902,22 +1902,22 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state)
> >   		intel_wait_for_pipe_off(old_crtc_state);
> >   }
> > -static unsigned int intel_tile_size(const struct drm_i915_private *dev_priv)
> > +static unsigned int intel_tile_size(const struct drm_i915_private *i915)
> >   {
> > -	return IS_GEN(dev_priv, 2) ? 2048 : 4096;
> > +	return IS_GEN(i915, 2) ? 2048 : 4096;
> >   }
> >   static unsigned int
> >   intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(fb->dev);
> > +	struct drm_i915_private *i915 = to_i915(fb->dev);
> >   	unsigned int cpp = fb->format->cpp[color_plane];
> >   	switch (fb->modifier) {
> >   	case DRM_FORMAT_MOD_LINEAR:
> > -		return intel_tile_size(dev_priv);
> > +		return intel_tile_size(i915);
> >   	case I915_FORMAT_MOD_X_TILED:
> > -		if (IS_GEN(dev_priv, 2))
> > +		if (IS_GEN(i915, 2))
> >   			return 128;
> >   		else
> >   			return 512;
> > @@ -1926,7 +1926,7 @@ intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane)
> >   			return 128;
> >   		/* fall through */
> >   	case I915_FORMAT_MOD_Y_TILED:
> > -		if (IS_GEN(dev_priv, 2) || HAS_128_BYTE_Y_TILING(dev_priv))
> > +		if (IS_GEN(i915, 2) || HAS_128_BYTE_Y_TILING(i915))
> >   			return 128;
> >   		else
> >   			return 512;
> > @@ -2017,26 +2017,26 @@ intel_fill_fb_ggtt_view(struct i915_ggtt_view *view,
> >   	}
> >   }
> > -static unsigned int intel_cursor_alignment(const struct drm_i915_private *dev_priv)
> > +static unsigned int intel_cursor_alignment(const struct drm_i915_private *i915)
> >   {
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		return 16 * 1024;
> > -	else if (IS_I85X(dev_priv))
> > +	else if (IS_I85X(i915))
> >   		return 256;
> > -	else if (IS_I845G(dev_priv) || IS_I865G(dev_priv))
> > +	else if (IS_I845G(i915) || IS_I865G(i915))
> >   		return 32;
> >   	else
> >   		return 4 * 1024;
> >   }
> > -static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_priv)
> > +static unsigned int intel_linear_alignment(const struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return 256 * 1024;
> > -	else if (IS_I965G(dev_priv) || IS_I965GM(dev_priv) ||
> > -		 IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_I965G(i915) || IS_I965GM(i915) ||
> > +		 IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		return 128 * 1024;
> > -	else if (INTEL_GEN(dev_priv) >= 4)
> > +	else if (INTEL_GEN(i915) >= 4)
> >   		return 4 * 1024;
> >   	else
> >   		return 0;
> > @@ -2045,7 +2045,7 @@ static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_pr
> >   static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
> >   					 int color_plane)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(fb->dev);
> > +	struct drm_i915_private *i915 = to_i915(fb->dev);
> >   	/* AUX_DIST needs only 4K alignment */
> >   	if (color_plane == 1)
> > @@ -2053,9 +2053,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
> >   	switch (fb->modifier) {
> >   	case DRM_FORMAT_MOD_LINEAR:
> > -		return intel_linear_alignment(dev_priv);
> > +		return intel_linear_alignment(i915);
> >   	case I915_FORMAT_MOD_X_TILED:
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			return 256 * 1024;
> >   		return 0;
> >   	case I915_FORMAT_MOD_Y_TILED_CCS:
> > @@ -2072,9 +2072,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb,
> >   static bool intel_plane_uses_fence(const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> > -	return INTEL_GEN(dev_priv) < 4 ||
> > +	return INTEL_GEN(i915) < 4 ||
> >   		(plane->has_fbc &&
> >   		 plane_state->view.type == I915_GGTT_VIEW_NORMAL);
> >   }
> > @@ -2086,7 +2086,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   			   unsigned long *out_flags)
> >   {
> >   	struct drm_device *dev = fb->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> >   	intel_wakeref_t wakeref;
> >   	struct i915_vma *vma;
> > @@ -2102,7 +2102,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   	 * we should always have valid PTE following the scanout preventing
> >   	 * the VT-d warning.
> >   	 */
> > -	if (intel_scanout_needs_vtd_wa(dev_priv) && alignment < 256 * 1024)
> > +	if (intel_scanout_needs_vtd_wa(i915) && alignment < 256 * 1024)
> >   		alignment = 256 * 1024;
> >   	/*
> > @@ -2112,10 +2112,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   	 * intel_runtime_pm_put(), so it is correct to wrap only the
> >   	 * pin/unpin/fence and not more.
> >   	 */
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	i915_gem_object_lock(obj);
> > -	atomic_inc(&dev_priv->gpu_error.pending_fb_pin);
> > +	atomic_inc(&i915->gpu_error.pending_fb_pin);
> >   	pinctl = 0;
> > @@ -2126,7 +2126,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   	 * complicated than this. For example, Cherryview appears quite
> >   	 * happy to scanout from anywhere within its global aperture.
> >   	 */
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		pinctl |= PIN_MAPPABLE;
> >   	vma = i915_gem_object_pin_to_display_plane(obj,
> > @@ -2154,7 +2154,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   		 * mode that matches the user configuration.
> >   		 */
> >   		ret = i915_vma_pin_fence(vma);
> > -		if (ret != 0 && INTEL_GEN(dev_priv) < 4) {
> > +		if (ret != 0 && INTEL_GEN(i915) < 4) {
> >   			i915_gem_object_unpin_from_display_plane(vma);
> >   			vma = ERR_PTR(ret);
> >   			goto err;
> > @@ -2166,10 +2166,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb,
> >   	i915_vma_get(vma);
> >   err:
> > -	atomic_dec(&dev_priv->gpu_error.pending_fb_pin);
> > +	atomic_dec(&i915->gpu_error.pending_fb_pin);
> >   	i915_gem_object_unlock(obj);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	return vma;
> >   }
> > @@ -2265,7 +2265,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y,
> >   				       unsigned int pitch,
> >   				       u32 old_offset, u32 new_offset)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(fb->dev);
> > +	struct drm_i915_private *i915 = to_i915(fb->dev);
> >   	unsigned int cpp = fb->format->cpp[color_plane];
> >   	WARN_ON(new_offset > old_offset);
> > @@ -2274,7 +2274,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y,
> >   		unsigned int tile_size, tile_width, tile_height;
> >   		unsigned int pitch_tiles;
> > -		tile_size = intel_tile_size(dev_priv);
> > +		tile_size = intel_tile_size(i915);
> >   		intel_tile_dims(fb, color_plane, &tile_width, &tile_height);
> >   		if (drm_rotation_90_or_270(rotation)) {
> > @@ -2326,7 +2326,7 @@ static u32 intel_plane_adjust_aligned_offset(int *x, int *y,
> >    * used. This is why the user has to pass in the pitch since it
> >    * is specified in the rotated orientation.
> >    */
> > -static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv,
> > +static u32 intel_compute_aligned_offset(struct drm_i915_private *i915,
> >   					int *x, int *y,
> >   					const struct drm_framebuffer *fb,
> >   					int color_plane,
> > @@ -2344,7 +2344,7 @@ static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv,
> >   		unsigned int tile_size, tile_width, tile_height;
> >   		unsigned int tile_rows, tiles, pitch_tiles;
> > -		tile_size = intel_tile_size(dev_priv);
> > +		tile_size = intel_tile_size(i915);
> >   		intel_tile_dims(fb, color_plane, &tile_width, &tile_height);
> >   		if (drm_rotation_90_or_270(rotation)) {
> > @@ -2382,18 +2382,18 @@ static u32 intel_plane_compute_aligned_offset(int *x, int *y,
> >   					      int color_plane)
> >   {
> >   	struct intel_plane *intel_plane = to_intel_plane(state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_plane->base.dev);
> >   	const struct drm_framebuffer *fb = state->base.fb;
> >   	unsigned int rotation = state->base.rotation;
> >   	int pitch = state->color_plane[color_plane].stride;
> >   	u32 alignment;
> >   	if (intel_plane->id == PLANE_CURSOR)
> > -		alignment = intel_cursor_alignment(dev_priv);
> > +		alignment = intel_cursor_alignment(i915);
> >   	else
> >   		alignment = intel_surf_alignment(fb, color_plane);
> > -	return intel_compute_aligned_offset(dev_priv, x, y, fb, color_plane,
> > +	return intel_compute_aligned_offset(i915, x, y, fb, color_plane,
> >   					    pitch, rotation, alignment);
> >   }
> > @@ -2402,11 +2402,11 @@ static int intel_fb_offset_to_xy(int *x, int *y,
> >   				 const struct drm_framebuffer *fb,
> >   				 int color_plane)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(fb->dev);
> > +	struct drm_i915_private *i915 = to_i915(fb->dev);
> >   	unsigned int height;
> >   	if (fb->modifier != DRM_FORMAT_MOD_LINEAR &&
> > -	    fb->offsets[color_plane] % intel_tile_size(dev_priv)) {
> > +	    fb->offsets[color_plane] % intel_tile_size(i915)) {
> >   		DRM_DEBUG_KMS("Misaligned offset 0x%08x for color plane %d\n",
> >   			      fb->offsets[color_plane], color_plane);
> >   		return -EINVAL;
> > @@ -2507,7 +2507,7 @@ bool is_ccs_modifier(u64 modifier)
> >   	       modifier == I915_FORMAT_MOD_Yf_TILED_CCS;
> >   }
> > -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> > +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915,
> >   			      u32 pixel_format, u64 modifier)
> >   {
> >   	struct intel_crtc *crtc;
> > @@ -2517,7 +2517,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> >   	 * We assume the primary plane for pipe A has
> >   	 * the highest stride limits of them all.
> >   	 */
> > -	crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_A);
> > +	crtc = intel_get_crtc_for_pipe(i915, PIPE_A);
> >   	plane = to_intel_plane(crtc->base.primary);
> >   	return plane->max_stride(plane, pixel_format, modifier,
> > @@ -2525,7 +2525,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> >   }
> >   static
> > -u32 intel_fb_max_stride(struct drm_i915_private *dev_priv,
> > +u32 intel_fb_max_stride(struct drm_i915_private *i915,
> >   			u32 pixel_format, u64 modifier)
> >   {
> >   	/*
> > @@ -2535,22 +2535,22 @@ u32 intel_fb_max_stride(struct drm_i915_private *dev_priv,
> >   	 * The new CCS hash mode makes remapping impossible
> >   	 */
> >   	if (!is_ccs_modifier(modifier)) {
> > -		if (INTEL_GEN(dev_priv) >= 7)
> > +		if (INTEL_GEN(i915) >= 7)
> >   			return 256*1024;
> > -		else if (INTEL_GEN(dev_priv) >= 4)
> > +		else if (INTEL_GEN(i915) >= 4)
> >   			return 128*1024;
> >   	}
> > -	return intel_plane_fb_max_stride(dev_priv, pixel_format, modifier);
> > +	return intel_plane_fb_max_stride(i915, pixel_format, modifier);
> >   }
> >   static u32
> >   intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(fb->dev);
> > +	struct drm_i915_private *i915 = to_i915(fb->dev);
> >   	if (fb->modifier == DRM_FORMAT_MOD_LINEAR) {
> > -		u32 max_stride = intel_plane_fb_max_stride(dev_priv,
> > +		u32 max_stride = intel_plane_fb_max_stride(i915,
> >   							   fb->format->format,
> >   							   fb->modifier);
> > @@ -2559,7 +2559,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> >   		 * we need the stride to be page aligned.
> >   		 */
> >   		if (fb->pitches[color_plane] > max_stride)
> > -			return intel_tile_size(dev_priv);
> > +			return intel_tile_size(i915);
> >   		else
> >   			return 64;
> >   	} else {
> > @@ -2570,7 +2570,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane)
> >   bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	int i;
> > @@ -2584,7 +2584,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
> >   	 * Would also need to deal with the fence POT alignment
> >   	 * and gen2 2KiB GTT tile size.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 4)
> > +	if (INTEL_GEN(i915) < 4)
> >   		return false;
> >   	/*
> > @@ -2596,7 +2596,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state)
> >   	/* Linear needs a page aligned stride for remapping */
> >   	if (fb->modifier == DRM_FORMAT_MOD_LINEAR) {
> > -		unsigned int alignment = intel_tile_size(dev_priv) - 1;
> > +		unsigned int alignment = intel_tile_size(i915) - 1;
> >   		for (i = 0; i < fb->format->num_planes; i++) {
> >   			if (fb->pitches[i] & alignment)
> > @@ -2636,7 +2636,7 @@ static bool intel_plane_needs_remap(const struct intel_plane_state *plane_state)
> >   }
> >   static int
> > -intel_fill_fb_info(struct drm_i915_private *dev_priv,
> > +intel_fill_fb_info(struct drm_i915_private *i915,
> >   		   struct drm_framebuffer *fb)
> >   {
> >   	struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb);
> > @@ -2645,7 +2645,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> >   	u32 gtt_offset_rotated = 0;
> >   	unsigned int max_size = 0;
> >   	int i, num_planes = fb->format->num_planes;
> > -	unsigned int tile_size = intel_tile_size(dev_priv);
> > +	unsigned int tile_size = intel_tile_size(i915);
> >   	for (i = 0; i < num_planes; i++) {
> >   		unsigned int width, height;
> > @@ -2719,7 +2719,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> >   		intel_fb->normal[i].x = x;
> >   		intel_fb->normal[i].y = y;
> > -		offset = intel_compute_aligned_offset(dev_priv, &x, &y, fb, i,
> > +		offset = intel_compute_aligned_offset(i915, &x, &y, fb, i,
> >   						      fb->pitches[i],
> >   						      DRM_MODE_ROTATE_0,
> >   						      tile_size);
> > @@ -2803,14 +2803,14 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
> >   static void
> >   intel_plane_remap_gtt(struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	struct drm_framebuffer *fb = plane_state->base.fb;
> >   	struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb);
> >   	struct intel_rotation_info *info = &plane_state->view.rotated;
> >   	unsigned int rotation = plane_state->base.rotation;
> >   	int i, num_planes = fb->format->num_planes;
> > -	unsigned int tile_size = intel_tile_size(dev_priv);
> > +	unsigned int tile_size = intel_tile_size(i915);
> >   	unsigned int src_x, src_y;
> >   	unsigned int src_w, src_h;
> >   	u32 gtt_offset = 0;
> > @@ -2860,7 +2860,7 @@ intel_plane_remap_gtt(struct intel_plane_state *plane_state)
> >   		x += intel_fb->normal[i].x;
> >   		y += intel_fb->normal[i].y;
> > -		offset = intel_compute_aligned_offset(dev_priv, &x, &y,
> > +		offset = intel_compute_aligned_offset(i915, &x, &y,
> >   						      fb, i, fb->pitches[i],
> >   						      DRM_MODE_ROTATE_0, tile_size);
> >   		offset /= tile_size;
> > @@ -3046,7 +3046,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> >   			      struct intel_initial_plane_config *plane_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_i915_gem_object *obj = NULL;
> >   	struct drm_mode_fb_cmd2 mode_cmd = { 0 };
> >   	struct drm_framebuffer *fb = &plane_config->fb->base;
> > @@ -3062,7 +3062,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> >   	/* If the FB is too big, just don't use it since fbdev is not very
> >   	 * important and we should probably use that space with FBC or other
> >   	 * features. */
> > -	if (size_aligned * 2 > dev_priv->stolen_usable_size)
> > +	if (size_aligned * 2 > i915->stolen_usable_size)
> >   		return false;
> >   	switch (fb->modifier) {
> > @@ -3077,7 +3077,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc,
> >   	}
> >   	mutex_lock(&dev->struct_mutex);
> > -	obj = i915_gem_object_create_stolen_for_preallocated(dev_priv,
> > +	obj = i915_gem_object_create_stolen_for_preallocated(i915,
> >   							     base_aligned,
> >   							     base_aligned,
> >   							     size_aligned);
> > @@ -3135,7 +3135,7 @@ intel_set_plane_visible(struct intel_crtc_state *crtc_state,
> >   static void fixup_active_planes(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct drm_plane *plane;
> >   	/*
> > @@ -3145,7 +3145,7 @@ static void fixup_active_planes(struct intel_crtc_state *crtc_state)
> >   	 */
> >   	crtc_state->active_planes = 0;
> > -	drm_for_each_plane_mask(plane, &dev_priv->drm,
> > +	drm_for_each_plane_mask(plane, &i915->drm,
> >   				crtc_state->base.plane_mask)
> >   		crtc_state->active_planes |= BIT(to_intel_plane(plane)->id);
> >   }
> > @@ -3177,7 +3177,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc,
> >   			     struct intel_initial_plane_config *plane_config)
> >   {
> >   	struct drm_device *dev = intel_crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_crtc *c;
> >   	struct drm_i915_gem_object *obj;
> >   	struct drm_plane *primary = intel_crtc->base.primary;
> > @@ -3272,7 +3272,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc,
> >   	intel_state->base.dst = drm_plane_state_dest(plane_state);
> >   	if (i915_gem_object_is_tiled(obj))
> > -		dev_priv->preserve_bios_swizzle = true;
> > +		i915->preserve_bios_swizzle = true;
> >   	plane_state->fb = fb;
> >   	plane_state->crtc = &intel_crtc->base;
> > @@ -3381,7 +3381,7 @@ static bool skl_check_main_ccs_coordinates(struct intel_plane_state *plane_state
> >   static int skl_check_main_surface(struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane_state->base.plane->dev);
> > +	struct drm_i915_private *i915 = to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> >   	int x = plane_state->base.src.x1 >> 16;
> > @@ -3392,9 +3392,9 @@ static int skl_check_main_surface(struct intel_plane_state *plane_state)
> >   	int max_height = 4096;
> >   	u32 alignment, offset, aux_offset = plane_state->color_plane[1].offset;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		max_width = icl_max_plane_width(fb, 0, rotation);
> > -	else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		max_width = glk_max_plane_width(fb, 0, rotation);
> >   	else
> >   		max_width = skl_max_plane_width(fb, 0, rotation);
> > @@ -3564,16 +3564,16 @@ i9xx_plane_max_stride(struct intel_plane *plane,
> >   		      u32 pixel_format, u64 modifier,
> >   		      unsigned int rotation)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> > -	if (!HAS_GMCH(dev_priv)) {
> > +	if (!HAS_GMCH(i915)) {
> >   		return 32*1024;
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > +	} else if (INTEL_GEN(i915) >= 4) {
> >   		if (modifier == I915_FORMAT_MOD_X_TILED)
> >   			return 16*1024;
> >   		else
> >   			return 32*1024;
> > -	} else if (INTEL_GEN(dev_priv) >= 3) {
> > +	} else if (INTEL_GEN(i915) >= 3) {
> >   		if (modifier == I915_FORMAT_MOD_X_TILED)
> >   			return 8*1024;
> >   		else
> > @@ -3589,7 +3589,7 @@ i9xx_plane_max_stride(struct intel_plane *plane,
> >   static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 dspcntr = 0;
> >   	if (crtc_state->gamma_enable)
> > @@ -3598,7 +3598,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   	if (crtc_state->csc_enable)
> >   		dspcntr |= DISPPLANE_PIPE_CSC_ENABLE;
> > -	if (INTEL_GEN(dev_priv) < 5)
> > +	if (INTEL_GEN(i915) < 5)
> >   		dspcntr |= DISPPLANE_SEL_PIPE(crtc->pipe);
> >   	return dspcntr;
> > @@ -3607,7 +3607,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   			  const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> > @@ -3615,8 +3615,8 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   	dspcntr = DISPLAY_PLANE_ENABLE;
> > -	if (IS_G4X(dev_priv) || IS_GEN(dev_priv, 5) ||
> > -	    IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> > +	if (IS_G4X(i915) || IS_GEN(i915, 5) ||
> > +	    IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> >   		dspcntr |= DISPPLANE_TRICKLE_FEED_DISABLE;
> >   	switch (fb->format->format) {
> > @@ -3646,7 +3646,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   		return 0;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4 &&
> > +	if (INTEL_GEN(i915) >= 4 &&
> >   	    fb->modifier == I915_FORMAT_MOD_X_TILED)
> >   		dspcntr |= DISPPLANE_TILED;
> > @@ -3661,7 +3661,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	int src_x, src_y;
> >   	u32 offset;
> > @@ -3679,7 +3679,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
> >   	intel_add_fb_offsets(&src_x, &src_y, plane_state, 0);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		offset = intel_plane_compute_aligned_offset(&src_x, &src_y,
> >   							    plane_state, 0);
> >   	else
> > @@ -3694,7 +3694,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state)
> >   			   (src_y << 16) - plane_state->base.src.y1);
> >   	/* HSW/BDW do this automagically in hardware */
> > -	if (!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)) {
> > +	if (!IS_HASWELL(i915) && !IS_BROADWELL(i915)) {
> >   		unsigned int rotation = plane_state->base.rotation;
> >   		int src_w = drm_rect_width(&plane_state->base.src) >> 16;
> >   		int src_h = drm_rect_height(&plane_state->base.src) >> 16;
> > @@ -3752,7 +3752,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   			      const struct intel_crtc_state *crtc_state,
> >   			      const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> >   	u32 linear_offset;
> >   	int x = plane_state->color_plane[0].x;
> > @@ -3765,16 +3765,16 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   	linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		dspaddr_offset = plane_state->color_plane[0].offset;
> >   	else
> >   		dspaddr_offset = linear_offset;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(DSPSTRIDE(i9xx_plane), plane_state->color_plane[0].stride);
> > -	if (INTEL_GEN(dev_priv) < 4) {
> > +	if (INTEL_GEN(i915) < 4) {
> >   		/* pipesrc and dspsize control the size that is scaled from,
> >   		 * which should always be the user's requested size.
> >   		 */
> > @@ -3782,7 +3782,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   		I915_WRITE_FW(DSPSIZE(i9xx_plane),
> >   			      ((crtc_state->pipe_src_h - 1) << 16) |
> >   			      (crtc_state->pipe_src_w - 1));
> > -	} else if (IS_CHERRYVIEW(dev_priv) && i9xx_plane == PLANE_B) {
> > +	} else if (IS_CHERRYVIEW(i915) && i9xx_plane == PLANE_B) {
> >   		I915_WRITE_FW(PRIMPOS(i9xx_plane), 0);
> >   		I915_WRITE_FW(PRIMSIZE(i9xx_plane),
> >   			      ((crtc_state->pipe_src_h - 1) << 16) |
> > @@ -3790,9 +3790,9 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   		I915_WRITE_FW(PRIMCNSTALPHA(i9xx_plane), 0);
> >   	}
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		I915_WRITE_FW(DSPOFFSET(i9xx_plane), (y << 16) | x);
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > +	} else if (INTEL_GEN(i915) >= 4) {
> >   		I915_WRITE_FW(DSPLINOFF(i9xx_plane), linear_offset);
> >   		I915_WRITE_FW(DSPTILEOFF(i9xx_plane), (y << 16) | x);
> >   	}
> > @@ -3803,7 +3803,7 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   	 * the control register just before the surface register.
> >   	 */
> >   	I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		I915_WRITE_FW(DSPSURF(i9xx_plane),
> >   			      intel_plane_ggtt_offset(plane_state) +
> >   			      dspaddr_offset);
> > @@ -3812,13 +3812,13 @@ static void i9xx_update_plane(struct intel_plane *plane,
> >   			      intel_plane_ggtt_offset(plane_state) +
> >   			      dspaddr_offset);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void i9xx_disable_plane(struct intel_plane *plane,
> >   			       const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> >   	unsigned long irqflags;
> >   	u32 dspcntr;
> > @@ -3835,21 +3835,21 @@ static void i9xx_disable_plane(struct intel_plane *plane,
> >   	 */
> >   	dspcntr = i9xx_plane_ctl_crtc(crtc_state);
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		I915_WRITE_FW(DSPSURF(i9xx_plane), 0);
> >   	else
> >   		I915_WRITE_FW(DSPADDR(i9xx_plane), 0);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> >   				    enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> >   	intel_wakeref_t wakeref;
> > @@ -3862,7 +3862,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> >   	 * display power wells.
> >   	 */
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -3870,13 +3870,13 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> >   	ret = val & DISPLAY_PLANE_ENABLE;
> > -	if (INTEL_GEN(dev_priv) >= 5)
> > +	if (INTEL_GEN(i915) >= 5)
> >   		*pipe = plane->pipe;
> >   	else
> >   		*pipe = (val & DISPPLANE_SEL_PIPE_MASK) >>
> >   			DISPPLANE_SEL_PIPE_SHIFT;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -3884,7 +3884,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane,
> >   static void skl_detach_scaler(struct intel_crtc *intel_crtc, int id)
> >   {
> >   	struct drm_device *dev = intel_crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	I915_WRITE(SKL_PS_CTRL(intel_crtc->pipe, id), 0);
> >   	I915_WRITE(SKL_PS_WIN_POS(intel_crtc->pipe, id), 0);
> > @@ -4091,10 +4091,10 @@ static u32 cnl_plane_ctl_flip(unsigned int reflect)
> >   u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	u32 plane_ctl = 0;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		return plane_ctl;
> >   	if (crtc_state->gamma_enable)
> > @@ -4109,7 +4109,7 @@ u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   		  const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> > @@ -4118,7 +4118,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   	plane_ctl = PLANE_CTL_ENABLE;
> > -	if (INTEL_GEN(dev_priv) < 10 && !IS_GEMINILAKE(dev_priv)) {
> > +	if (INTEL_GEN(i915) < 10 && !IS_GEMINILAKE(i915)) {
> >   		plane_ctl |= skl_plane_ctl_alpha(plane_state);
> >   		plane_ctl |= PLANE_CTL_PLANE_GAMMA_DISABLE;
> > @@ -4133,7 +4133,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   	plane_ctl |= skl_plane_ctl_tiling(fb->modifier);
> >   	plane_ctl |= skl_plane_ctl_rotate(rotation & DRM_MODE_ROTATE_MASK);
> > -	if (INTEL_GEN(dev_priv) >= 10)
> > +	if (INTEL_GEN(i915) >= 10)
> >   		plane_ctl |= cnl_plane_ctl_flip(rotation &
> >   						DRM_MODE_REFLECT_MASK);
> > @@ -4147,10 +4147,10 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state,
> >   u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	u32 plane_color_ctl = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return plane_color_ctl;
> >   	if (crtc_state->gamma_enable)
> > @@ -4165,7 +4165,7 @@ u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state,
> >   			const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > @@ -4174,7 +4174,7 @@ u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state,
> >   	plane_color_ctl |= PLANE_COLOR_PLANE_GAMMA_DISABLE;
> >   	plane_color_ctl |= glk_plane_color_ctl_alpha(plane_state);
> > -	if (fb->format->is_yuv && !icl_is_hdr_plane(dev_priv, plane->id)) {
> > +	if (fb->format->is_yuv && !icl_is_hdr_plane(i915, plane->id)) {
> >   		if (plane_state->base.color_encoding == DRM_COLOR_YCBCR_BT709)
> >   			plane_color_ctl |= PLANE_COLOR_CSC_MODE_YUV709_TO_RGB709;
> >   		else
> > @@ -4228,31 +4228,31 @@ __intel_display_resume(struct drm_device *dev,
> >   	return ret;
> >   }
> > -static bool gpu_reset_clobbers_display(struct drm_i915_private *dev_priv)
> > +static bool gpu_reset_clobbers_display(struct drm_i915_private *i915)
> >   {
> > -	return (INTEL_INFO(dev_priv)->gpu_reset_clobbers_display &&
> > -		intel_has_gpu_reset(dev_priv));
> > +	return (INTEL_INFO(i915)->gpu_reset_clobbers_display &&
> > +		intel_has_gpu_reset(i915));
> >   }
> > -void intel_prepare_reset(struct drm_i915_private *dev_priv)
> > +void intel_prepare_reset(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > -	struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx;
> > +	struct drm_device *dev = &i915->drm;
> > +	struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx;
> >   	struct drm_atomic_state *state;
> >   	int ret;
> >   	/* reset doesn't touch the display */
> >   	if (!i915_modparams.force_reset_modeset_test &&
> > -	    !gpu_reset_clobbers_display(dev_priv))
> > +	    !gpu_reset_clobbers_display(i915))
> >   		return;
> >   	/* We have a modeset vs reset deadlock, defensively unbreak it. */
> > -	set_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags);
> > -	wake_up_all(&dev_priv->gpu_error.wait_queue);
> > +	set_bit(I915_RESET_MODESET, &i915->gpu_error.flags);
> > +	wake_up_all(&i915->gpu_error.wait_queue);
> > -	if (atomic_read(&dev_priv->gpu_error.pending_fb_pin)) {
> > +	if (atomic_read(&i915->gpu_error.pending_fb_pin)) {
> >   		DRM_DEBUG_KMS("Modeset potentially stuck, unbreaking through wedging\n");
> > -		i915_gem_set_wedged(dev_priv);
> > +		i915_gem_set_wedged(i915);
> >   	}
> >   	/*
> > @@ -4286,27 +4286,27 @@ void intel_prepare_reset(struct drm_i915_private *dev_priv)
> >   		return;
> >   	}
> > -	dev_priv->modeset_restore_state = state;
> > +	i915->modeset_restore_state = state;
> >   	state->acquire_ctx = ctx;
> >   }
> > -void intel_finish_reset(struct drm_i915_private *dev_priv)
> > +void intel_finish_reset(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > -	struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx;
> > +	struct drm_device *dev = &i915->drm;
> > +	struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx;
> >   	struct drm_atomic_state *state;
> >   	int ret;
> >   	/* reset doesn't touch the display */
> > -	if (!test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags))
> > +	if (!test_bit(I915_RESET_MODESET, &i915->gpu_error.flags))
> >   		return;
> > -	state = fetch_and_zero(&dev_priv->modeset_restore_state);
> > +	state = fetch_and_zero(&i915->modeset_restore_state);
> >   	if (!state)
> >   		goto unlock;
> >   	/* reset doesn't touch the display */
> > -	if (!gpu_reset_clobbers_display(dev_priv)) {
> > +	if (!gpu_reset_clobbers_display(i915)) {
> >   		/* for testing only restore the display */
> >   		ret = __intel_display_resume(dev, state, ctx);
> >   		if (ret)
> > @@ -4316,20 +4316,20 @@ void intel_finish_reset(struct drm_i915_private *dev_priv)
> >   		 * The display has been reset as well,
> >   		 * so need a full re-initialization.
> >   		 */
> > -		intel_pps_unlock_regs_wa(dev_priv);
> > +		intel_pps_unlock_regs_wa(i915);
> >   		intel_modeset_init_hw(dev);
> > -		intel_init_clock_gating(dev_priv);
> > +		intel_init_clock_gating(i915);
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		if (dev_priv->display.hpd_irq_setup)
> > -			dev_priv->display.hpd_irq_setup(dev_priv);
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +		spin_lock_irq(&i915->irq_lock);
> > +		if (i915->display.hpd_irq_setup)
> > +			i915->display.hpd_irq_setup(i915);
> > +		spin_unlock_irq(&i915->irq_lock);
> >   		ret = __intel_display_resume(dev, state, ctx);
> >   		if (ret)
> >   			DRM_ERROR("Restoring old state failed with %i\n", ret);
> > -		intel_hpd_init(dev_priv);
> > +		intel_hpd_init(i915);
> >   	}
> >   	drm_atomic_state_put(state);
> > @@ -4338,12 +4338,12 @@ void intel_finish_reset(struct drm_i915_private *dev_priv)
> >   	drm_modeset_acquire_fini(ctx);
> >   	mutex_unlock(&dev->mode_config.mutex);
> > -	clear_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags);
> > +	clear_bit(I915_RESET_MODESET, &i915->gpu_error.flags);
> >   }
> >   static void icl_set_pipe_chicken(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 tmp;
> > @@ -4369,7 +4369,7 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta
> >   				     const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	/* drm_atomic_helper_update_legacy_modeset_state might not be called. */
> >   	crtc->base.mode = new_crtc_state->base.mode;
> > @@ -4388,26 +4388,26 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta
> >   		   (new_crtc_state->pipe_src_h - 1));
> >   	/* on skylake this is done by detaching scalers */
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		skl_detach_scalers(new_crtc_state);
> >   		if (new_crtc_state->pch_pfit.enabled)
> >   			skylake_pfit_enable(new_crtc_state);
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		if (new_crtc_state->pch_pfit.enabled)
> >   			ironlake_pfit_enable(new_crtc_state);
> >   		else if (old_crtc_state->pch_pfit.enabled)
> >   			ironlake_pfit_disable(old_crtc_state);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_set_pipe_chicken(crtc);
> >   }
> >   static void intel_fdi_normal_train(struct intel_crtc *crtc)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp;
> > @@ -4415,7 +4415,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
> >   	/* enable normal train */
> >   	reg = FDI_TX_CTL(pipe);
> >   	temp = I915_READ(reg);
> > -	if (IS_IVYBRIDGE(dev_priv)) {
> > +	if (IS_IVYBRIDGE(i915)) {
> >   		temp &= ~FDI_LINK_TRAIN_NONE_IVB;
> >   		temp |= FDI_LINK_TRAIN_NONE_IVB | FDI_TX_ENHANCE_FRAME_ENABLE;
> >   	} else {
> > @@ -4426,7 +4426,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
> >   	reg = FDI_RX_CTL(pipe);
> >   	temp = I915_READ(reg);
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> >   		temp |= FDI_LINK_TRAIN_NORMAL_CPT;
> >   	} else {
> > @@ -4440,7 +4440,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc)
> >   	udelay(1000);
> >   	/* IVB wants error correction enabled */
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		I915_WRITE(reg, I915_READ(reg) | FDI_FS_ERRC_ENABLE |
> >   			   FDI_FE_ERRC_ENABLE);
> >   }
> > @@ -4450,13 +4450,13 @@ static void ironlake_fdi_link_train(struct intel_crtc *crtc,
> >   				    const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp, tries;
> >   	/* FDI needs bits from pipe first */
> > -	assert_pipe_enabled(dev_priv, pipe);
> > +	assert_pipe_enabled(i915, pipe);
> >   	/* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit
> >   	   for train result */
> > @@ -4551,7 +4551,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> >   				const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp, i, retry;
> > @@ -4584,7 +4584,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> >   	reg = FDI_RX_CTL(pipe);
> >   	temp = I915_READ(reg);
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> >   		temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
> >   	} else {
> > @@ -4628,7 +4628,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> >   	temp = I915_READ(reg);
> >   	temp &= ~FDI_LINK_TRAIN_NONE;
> >   	temp |= FDI_LINK_TRAIN_PATTERN_2;
> > -	if (IS_GEN(dev_priv, 6)) {
> > +	if (IS_GEN(i915, 6)) {
> >   		temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK;
> >   		/* SNB-B */
> >   		temp |= FDI_LINK_TRAIN_400MV_0DB_SNB_B;
> > @@ -4637,7 +4637,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc,
> >   	reg = FDI_RX_CTL(pipe);
> >   	temp = I915_READ(reg);
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> >   		temp |= FDI_LINK_TRAIN_PATTERN_2_CPT;
> >   	} else {
> > @@ -4684,7 +4684,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc,
> >   				      const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp, i, j;
> > @@ -4802,7 +4802,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc,
> >   static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	int pipe = intel_crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp;
> > @@ -4839,7 +4839,7 @@ static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state)
> >   static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc)
> >   {
> >   	struct drm_device *dev = intel_crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = intel_crtc->pipe;
> >   	i915_reg_t reg;
> >   	u32 temp;
> > @@ -4869,7 +4869,7 @@ static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc)
> >   static void ironlake_fdi_disable(struct drm_crtc *crtc)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> >   	i915_reg_t reg;
> > @@ -4891,7 +4891,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
> >   	udelay(100);
> >   	/* Ironlake workaround, disable clock pointer after downing FDI */
> > -	if (HAS_PCH_IBX(dev_priv))
> > +	if (HAS_PCH_IBX(i915))
> >   		I915_WRITE(FDI_RX_CHICKEN(pipe), FDI_RX_PHASE_SYNC_POINTER_OVR);
> >   	/* still set train pattern 1 */
> > @@ -4903,7 +4903,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
> >   	reg = FDI_RX_CTL(pipe);
> >   	temp = I915_READ(reg);
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
> >   		temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
> >   	} else {
> > @@ -4919,12 +4919,12 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc)
> >   	udelay(100);
> >   }
> > -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv)
> > +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915)
> >   {
> >   	struct drm_crtc *crtc;
> >   	bool cleanup_done;
> > -	drm_for_each_crtc(crtc, &dev_priv->drm) {
> > +	drm_for_each_crtc(crtc, &i915->drm) {
> >   		struct drm_crtc_commit *commit;
> >   		spin_lock(&crtc->commit_lock);
> >   		commit = list_first_entry_or_null(&crtc->commit_list,
> > @@ -4944,31 +4944,31 @@ bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv)
> >   	return false;
> >   }
> > -void lpt_disable_iclkip(struct drm_i915_private *dev_priv)
> > +void lpt_disable_iclkip(struct drm_i915_private *i915)
> >   {
> >   	u32 temp;
> >   	I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_GATE);
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> >   	temp |= SBI_SSCCTL_DISABLE;
> > -	intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK);
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   }
> >   /* Program iCLKIP clock to the desired frequency */
> >   static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int clock = crtc_state->base.adjusted_mode.crtc_clock;
> >   	u32 divsel, phaseinc, auxdiv, phasedir = 0;
> >   	u32 temp;
> > -	lpt_disable_iclkip(dev_priv);
> > +	lpt_disable_iclkip(i915);
> >   	/* The iCLK virtual clock root frequency is in MHz,
> >   	 * but the adjusted_mode->crtc_clock in in KHz. To get the
> > @@ -5007,30 +5007,30 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> >   			phasedir,
> >   			phaseinc);
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> >   	/* Program SSCDIVINTPHASE6 */
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> >   	temp &= ~SBI_SSCDIVINTPHASE_DIVSEL_MASK;
> >   	temp |= SBI_SSCDIVINTPHASE_DIVSEL(divsel);
> >   	temp &= ~SBI_SSCDIVINTPHASE_INCVAL_MASK;
> >   	temp |= SBI_SSCDIVINTPHASE_INCVAL(phaseinc);
> >   	temp |= SBI_SSCDIVINTPHASE_DIR(phasedir);
> >   	temp |= SBI_SSCDIVINTPHASE_PROPAGATE;
> > -	intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK);
> >   	/* Program SSCAUXDIV */
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK);
> >   	temp &= ~SBI_SSCAUXDIV_FINALDIV2SEL(1);
> >   	temp |= SBI_SSCAUXDIV_FINALDIV2SEL(auxdiv);
> > -	intel_sbi_write(dev_priv, SBI_SSCAUXDIV6, temp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCAUXDIV6, temp, SBI_ICLK);
> >   	/* Enable modulator and associated divider */
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> >   	temp &= ~SBI_SSCCTL_DISABLE;
> > -	intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK);
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   	/* Wait for initialization time */
> >   	udelay(24);
> > @@ -5038,7 +5038,7 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state)
> >   	I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_UNGATE);
> >   }
> > -int lpt_get_iclkip(struct drm_i915_private *dev_priv)
> > +int lpt_get_iclkip(struct drm_i915_private *i915)
> >   {
> >   	u32 divsel, phaseinc, auxdiv;
> >   	u32 iclk_virtual_root_freq = 172800 * 1000;
> > @@ -5049,25 +5049,25 @@ int lpt_get_iclkip(struct drm_i915_private *dev_priv)
> >   	if ((I915_READ(PIXCLK_GATE) & PIXCLK_GATE_UNGATE) == 0)
> >   		return 0;
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK);
> >   	if (temp & SBI_SSCCTL_DISABLE) {
> > -		mutex_unlock(&dev_priv->sb_lock);
> > +		mutex_unlock(&i915->sb_lock);
> >   		return 0;
> >   	}
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK);
> >   	divsel = (temp & SBI_SSCDIVINTPHASE_DIVSEL_MASK) >>
> >   		SBI_SSCDIVINTPHASE_DIVSEL_SHIFT;
> >   	phaseinc = (temp & SBI_SSCDIVINTPHASE_INCVAL_MASK) >>
> >   		SBI_SSCDIVINTPHASE_INCVAL_SHIFT;
> > -	temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK);
> > +	temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK);
> >   	auxdiv = (temp & SBI_SSCAUXDIV_FINALDIV2SEL_MASK) >>
> >   		SBI_SSCAUXDIV_FINALDIV2SEL_SHIFT;
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   	desired_divisor = (divsel + 2) * iclk_pi_range + phaseinc;
> > @@ -5079,7 +5079,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c
> >   						enum pipe pch_transcoder)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	I915_WRITE(PCH_TRANS_HTOTAL(pch_transcoder),
> > @@ -5099,7 +5099,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c
> >   		   I915_READ(VSYNCSHIFT(cpu_transcoder)));
> >   }
> > -static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool enable)
> > +static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *i915, bool enable)
> >   {
> >   	u32 temp;
> > @@ -5122,20 +5122,20 @@ static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool e
> >   static void ivybridge_update_fdi_bc_bifurcation(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	switch (crtc->pipe) {
> >   	case PIPE_A:
> >   		break;
> >   	case PIPE_B:
> >   		if (crtc_state->fdi_lanes > 2)
> > -			cpt_set_fdi_bc_bifurcation(dev_priv, false);
> > +			cpt_set_fdi_bc_bifurcation(i915, false);
> >   		else
> > -			cpt_set_fdi_bc_bifurcation(dev_priv, true);
> > +			cpt_set_fdi_bc_bifurcation(i915, true);
> >   		break;
> >   	case PIPE_C:
> > -		cpt_set_fdi_bc_bifurcation(dev_priv, true);
> > +		cpt_set_fdi_bc_bifurcation(i915, true);
> >   		break;
> >   	default:
> > @@ -5185,13 +5185,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = crtc->pipe;
> >   	u32 temp;
> > -	assert_pch_transcoder_disabled(dev_priv, pipe);
> > +	assert_pch_transcoder_disabled(i915, pipe);
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		ivybridge_update_fdi_bc_bifurcation(crtc_state);
> >   	/* Write the TU size bits before fdi link training, so that error
> > @@ -5200,18 +5200,18 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> >   		   I915_READ(PIPE_DATA_M1(pipe)) & TU_SIZE_MASK);
> >   	/* For PCH output, training FDI link */
> > -	dev_priv->display.fdi_link_train(crtc, crtc_state);
> > +	i915->display.fdi_link_train(crtc, crtc_state);
> >   	/* We need to program the right clock selection before writing the pixel
> >   	 * mutliplier into the DPLL. */
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		u32 sel;
> >   		temp = I915_READ(PCH_DPLL_SEL);
> >   		temp |= TRANS_DPLL_ENABLE(pipe);
> >   		sel = TRANS_DPLLB_SEL(pipe);
> >   		if (crtc_state->shared_dpll ==
> > -		    intel_get_shared_dpll_by_id(dev_priv, DPLL_ID_PCH_PLL_B))
> > +		    intel_get_shared_dpll_by_id(i915, DPLL_ID_PCH_PLL_B))
> >   			temp |= sel;
> >   		else
> >   			temp &= ~sel;
> > @@ -5228,13 +5228,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state,
> >   	intel_enable_shared_dpll(crtc_state);
> >   	/* set transcoder timing, panel must allow it */
> > -	assert_panel_unlocked(dev_priv, pipe);
> > +	assert_panel_unlocked(i915, pipe);
> >   	ironlake_pch_transcoder_set_timings(crtc_state, pipe);
> >   	intel_fdi_normal_train(crtc);
> >   	/* For PCH DP, enable TRANS_DP_CTL */
> > -	if (HAS_PCH_CPT(dev_priv) &&
> > +	if (HAS_PCH_CPT(i915) &&
> >   	    intel_crtc_has_dp_encoder(crtc_state)) {
> >   		const struct drm_display_mode *adjusted_mode =
> >   			&crtc_state->base.adjusted_mode;
> > @@ -5268,22 +5268,22 @@ static void lpt_pch_enable(const struct intel_atomic_state *state,
> >   			   const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> > -	assert_pch_transcoder_disabled(dev_priv, PIPE_A);
> > +	assert_pch_transcoder_disabled(i915, PIPE_A);
> >   	lpt_program_iclkip(crtc_state);
> >   	/* Set transcoder timing. */
> >   	ironlake_pch_transcoder_set_timings(crtc_state, PIPE_A);
> > -	lpt_enable_pch_transcoder(dev_priv, cpu_transcoder);
> > +	lpt_enable_pch_transcoder(i915, cpu_transcoder);
> >   }
> >   static void cpt_verify_modeset(struct drm_device *dev, int pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	i915_reg_t dslreg = PIPEDSL(pipe);
> >   	u32 temp;
> > @@ -5385,7 +5385,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> >   		&crtc_state->scaler_state;
> >   	struct intel_crtc *intel_crtc =
> >   		to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	const struct drm_display_mode *adjusted_mode =
> >   		&crtc_state->base.adjusted_mode;
> > @@ -5403,7 +5403,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> >   	 * Once NV12 is enabled, handle it here while allocating scaler
> >   	 * for NV12.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 9 && crtc_state->base.enable &&
> > +	if (INTEL_GEN(i915) >= 9 && crtc_state->base.enable &&
> >   	    need_scaler && adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE) {
> >   		DRM_DEBUG_KMS("Pipe/Plane scaling not supported with IF-ID mode\n");
> >   		return -EINVAL;
> > @@ -5442,10 +5442,10 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach,
> >   	/* range checks */
> >   	if (src_w < SKL_MIN_SRC_W || src_h < SKL_MIN_SRC_H ||
> >   	    dst_w < SKL_MIN_DST_W || dst_h < SKL_MIN_DST_H ||
> > -	    (INTEL_GEN(dev_priv) >= 11 &&
> > +	    (INTEL_GEN(i915) >= 11 &&
> >   	     (src_w > ICL_MAX_SRC_W || src_h > ICL_MAX_SRC_H ||
> >   	      dst_w > ICL_MAX_DST_W || dst_h > ICL_MAX_DST_H)) ||
> > -	    (INTEL_GEN(dev_priv) < 11 &&
> > +	    (INTEL_GEN(i915) < 11 &&
> >   	     (src_w > SKL_MAX_SRC_W || src_h > SKL_MAX_SRC_H ||
> >   	      dst_w > SKL_MAX_DST_W || dst_h > SKL_MAX_DST_H)))	{
> >   		DRM_DEBUG_KMS("scaler_user index %u.%u: src %ux%u dst %ux%u "
> > @@ -5502,14 +5502,14 @@ static int skl_update_scaler_plane(struct intel_crtc_state *crtc_state,
> >   {
> >   	struct intel_plane *intel_plane =
> >   		to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_plane->base.dev);
> >   	struct drm_framebuffer *fb = plane_state->base.fb;
> >   	int ret;
> >   	bool force_detach = !fb || !plane_state->base.visible;
> >   	bool need_scaler = false;
> >   	/* Pre-gen11 and SDR planes always need a scaler for planar formats. */
> > -	if (!icl_is_hdr_plane(dev_priv, intel_plane->id) &&
> > +	if (!icl_is_hdr_plane(i915, intel_plane->id) &&
> >   	    fb && is_planar_yuv_format(fb->format->format))
> >   		need_scaler = true;
> > @@ -5582,7 +5582,7 @@ static void skylake_scaler_disable(struct intel_crtc *crtc)
> >   static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	const struct intel_crtc_scaler_state *scaler_state =
> >   		&crtc_state->scaler_state;
> > @@ -5619,7 +5619,7 @@ static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int pipe = crtc->pipe;
> >   	if (crtc_state->pch_pfit.enabled) {
> > @@ -5627,7 +5627,7 @@ static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   		 * as some pre-programmed values are broken,
> >   		 * e.g. x201.
> >   		 */
> > -		if (IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv))
> > +		if (IS_IVYBRIDGE(i915) || IS_HASWELL(i915))
> >   			I915_WRITE(PF_CTL(pipe), PF_ENABLE | PF_FILTER_MED_3x3 |
> >   						 PF_PIPE_SEL_IVB(pipe));
> >   		else
> > @@ -5641,7 +5641,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	if (!crtc_state->ips_enabled)
> >   		return;
> > @@ -5653,8 +5653,8 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> >   	 */
> >   	WARN_ON(!(crtc_state->active_planes & ~BIT(PLANE_CURSOR)));
> > -	if (IS_BROADWELL(dev_priv)) {
> > -		WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL,
> > +	if (IS_BROADWELL(i915)) {
> > +		WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL,
> >   						IPS_ENABLE | IPS_PCODE_CONTROL));
> >   		/* Quoting Art Runyan: "its not safe to expect any particular
> >   		 * value in IPS_CTL bit 31 after enabling IPS through the
> > @@ -5668,7 +5668,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state)
> >   		 * and don't wait for vblanks until the end of crtc_enable, then
> >   		 * the HW state readout code will complain that the expected
> >   		 * IPS_CTL value is not the one we read. */
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    IPS_CTL, IPS_ENABLE, IPS_ENABLE,
> >   					    50))
> >   			DRM_ERROR("Timed out waiting for IPS enable\n");
> > @@ -5679,19 +5679,19 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	if (!crtc_state->ips_enabled)
> >   		return;
> > -	if (IS_BROADWELL(dev_priv)) {
> > -		WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL, 0));
> > +	if (IS_BROADWELL(i915)) {
> > +		WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL, 0));
> >   		/*
> >   		 * Wait for PCODE to finish disabling IPS. The BSpec specified
> >   		 * 42ms timeout value leads to occasional timeouts so use 100ms
> >   		 * instead.
> >   		 */
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    IPS_CTL, IPS_ENABLE, 0,
> >   					    100))
> >   			DRM_ERROR("Timed out waiting for IPS disable\n");
> > @@ -5701,7 +5701,7 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state)
> >   	}
> >   	/* We need to wait for a vblank before we can disable the plane. */
> > -	intel_wait_for_vblank(dev_priv, crtc->pipe);
> > +	intel_wait_for_vblank(i915, crtc->pipe);
> >   }
> >   static void intel_crtc_dpms_overlay_disable(struct intel_crtc *intel_crtc)
> > @@ -5735,7 +5735,7 @@ intel_post_enable_primary(struct drm_crtc *crtc,
> >   			  const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> > @@ -5746,12 +5746,12 @@ intel_post_enable_primary(struct drm_crtc *crtc,
> >   	 * FIXME: Need to fix the logic to work when we turn off all planes
> >   	 * but leave the pipe running.
> >   	 */
> > -	if (IS_GEN(dev_priv, 2))
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	if (IS_GEN(i915, 2))
> > +		intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> >   	/* Underruns don't always raise interrupts, so check manually. */
> > -	intel_check_cpu_fifo_underruns(dev_priv);
> > -	intel_check_pch_fifo_underruns(dev_priv);
> > +	intel_check_cpu_fifo_underruns(i915);
> > +	intel_check_pch_fifo_underruns(i915);
> >   }
> >   /* FIXME get rid of this and use pre_plane_update */
> > @@ -5759,7 +5759,7 @@ static void
> >   intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> > @@ -5767,8 +5767,8 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> >   	 * Gen2 reports pipe underruns whenever all planes are disabled.
> >   	 * So disable underrun reporting before all the planes get disabled.
> >   	 */
> > -	if (IS_GEN(dev_priv, 2))
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> > +	if (IS_GEN(i915, 2))
> > +		intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> >   	hsw_disable_ips(to_intel_crtc_state(crtc->state));
> > @@ -5781,16 +5781,16 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc)
> >   	 * event which is after the vblank start event, so we need to have a
> >   	 * wait-for-vblank between disabling the plane and the pipe.
> >   	 */
> > -	if (HAS_GMCH(dev_priv) &&
> > -	    intel_set_memory_cxsr(dev_priv, false))
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +	if (HAS_GMCH(i915) &&
> > +	    intel_set_memory_cxsr(i915, false))
> > +		intel_wait_for_vblank(i915, pipe);
> >   }
> >   static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_state,
> >   				       const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	if (!old_crtc_state->ips_enabled)
> >   		return false;
> > @@ -5804,7 +5804,7 @@ static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_s
> >   	 *
> >   	 * Disable IPS before we program the LUT.
> >   	 */
> > -	if (IS_HASWELL(dev_priv) &&
> > +	if (IS_HASWELL(i915) &&
> >   	    (new_crtc_state->base.color_mgmt_changed ||
> >   	     new_crtc_state->update_pipe) &&
> >   	    new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT)
> > @@ -5817,7 +5817,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> >   				       const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	if (!new_crtc_state->ips_enabled)
> >   		return false;
> > @@ -5831,7 +5831,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> >   	 *
> >   	 * Re-enable IPS after the LUT has been programmed.
> >   	 */
> > -	if (IS_HASWELL(dev_priv) &&
> > +	if (IS_HASWELL(i915) &&
> >   	    (new_crtc_state->base.color_mgmt_changed ||
> >   	     new_crtc_state->update_pipe) &&
> >   	    new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT)
> > @@ -5848,24 +5848,24 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s
> >   	return !old_crtc_state->ips_enabled;
> >   }
> > -static bool needs_nv12_wa(struct drm_i915_private *dev_priv,
> > +static bool needs_nv12_wa(struct drm_i915_private *i915,
> >   			  const struct intel_crtc_state *crtc_state)
> >   {
> >   	if (!crtc_state->nv12_planes)
> >   		return false;
> >   	/* WA Display #0827: Gen9:all */
> > -	if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915))
> >   		return true;
> >   	return false;
> >   }
> > -static bool needs_scalerclk_wa(struct drm_i915_private *dev_priv,
> > +static bool needs_scalerclk_wa(struct drm_i915_private *i915,
> >   			       const struct intel_crtc_state *crtc_state)
> >   {
> >   	/* Wa_2006604312:icl */
> > -	if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(dev_priv))
> > +	if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(i915))
> >   		return true;
> >   	return false;
> > @@ -5875,7 +5875,7 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_atomic_state *old_state = old_crtc_state->base.state;
> >   	struct intel_crtc_state *pipe_config =
> >   		intel_atomic_get_new_crtc_state(to_intel_atomic_state(old_state),
> > @@ -5904,13 +5904,13 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state)
> >   			intel_post_enable_primary(&crtc->base, pipe_config);
> >   	}
> > -	if (needs_nv12_wa(dev_priv, old_crtc_state) &&
> > -	    !needs_nv12_wa(dev_priv, pipe_config))
> > -		skl_wa_827(dev_priv, crtc->pipe, false);
> > +	if (needs_nv12_wa(i915, old_crtc_state) &&
> > +	    !needs_nv12_wa(i915, pipe_config))
> > +		skl_wa_827(i915, crtc->pipe, false);
> > -	if (needs_scalerclk_wa(dev_priv, old_crtc_state) &&
> > -	    !needs_scalerclk_wa(dev_priv, pipe_config))
> > -		icl_wa_scalerclkgating(dev_priv, crtc->pipe, false);
> > +	if (needs_scalerclk_wa(i915, old_crtc_state) &&
> > +	    !needs_scalerclk_wa(i915, pipe_config))
> > +		icl_wa_scalerclkgating(i915, crtc->pipe, false);
> >   }
> >   static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> > @@ -5918,7 +5918,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_atomic_state *old_state = old_crtc_state->base.state;
> >   	struct drm_plane *primary = crtc->base.primary;
> >   	struct drm_plane_state *old_primary_state =
> > @@ -5940,20 +5940,20 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   		 * Gen2 reports pipe underruns whenever all planes are disabled.
> >   		 * So disable underrun reporting before all the planes get disabled.
> >   		 */
> > -		if (IS_GEN(dev_priv, 2) && old_primary_state->visible &&
> > +		if (IS_GEN(i915, 2) && old_primary_state->visible &&
> >   		    (modeset || !new_primary_state->base.visible))
> > -			intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false);
> > +			intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false);
> >   	}
> >   	/* Display WA 827 */
> > -	if (!needs_nv12_wa(dev_priv, old_crtc_state) &&
> > -	    needs_nv12_wa(dev_priv, pipe_config))
> > -		skl_wa_827(dev_priv, crtc->pipe, true);
> > +	if (!needs_nv12_wa(i915, old_crtc_state) &&
> > +	    needs_nv12_wa(i915, pipe_config))
> > +		skl_wa_827(i915, crtc->pipe, true);
> >   	/* Wa_2006604312:icl */
> > -	if (!needs_scalerclk_wa(dev_priv, old_crtc_state) &&
> > -	    needs_scalerclk_wa(dev_priv, pipe_config))
> > -		icl_wa_scalerclkgating(dev_priv, crtc->pipe, true);
> > +	if (!needs_scalerclk_wa(i915, old_crtc_state) &&
> > +	    needs_scalerclk_wa(i915, pipe_config))
> > +		icl_wa_scalerclkgating(i915, crtc->pipe, true);
> >   	/*
> >   	 * Vblank time updates from the shadow to live plane control register
> > @@ -5964,9 +5964,9 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   	 * event which is after the vblank start event, so we need to have a
> >   	 * wait-for-vblank between disabling the plane and the pipe.
> >   	 */
> > -	if (HAS_GMCH(dev_priv) && old_crtc_state->base.active &&
> > -	    pipe_config->disable_cxsr && intel_set_memory_cxsr(dev_priv, false))
> > -		intel_wait_for_vblank(dev_priv, crtc->pipe);
> > +	if (HAS_GMCH(i915) && old_crtc_state->base.active &&
> > +	    pipe_config->disable_cxsr && intel_set_memory_cxsr(i915, false))
> > +		intel_wait_for_vblank(i915, crtc->pipe);
> >   	/*
> >   	 * IVB workaround: must disable low power watermarks for at least
> > @@ -5977,7 +5977,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   	 */
> >   	if (pipe_config->disable_lp_wm && ilk_disable_lp_wm(dev) &&
> >   	    old_crtc_state->base.active)
> > -		intel_wait_for_vblank(dev_priv, crtc->pipe);
> > +		intel_wait_for_vblank(i915, crtc->pipe);
> >   	/*
> >   	 * If we're doing a modeset, we're done.  No need to do any pre-vblank
> > @@ -6000,8 +6000,8 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   	 * we'll continue to update watermarks the old way, if flags tell
> >   	 * us to.
> >   	 */
> > -	if (dev_priv->display.initial_watermarks != NULL)
> > -		dev_priv->display.initial_watermarks(old_intel_state,
> > +	if (i915->display.initial_watermarks != NULL)
> > +		i915->display.initial_watermarks(old_intel_state,
> >   						     pipe_config);
> >   	else if (pipe_config->update_wm_pre)
> >   		intel_update_watermarks(crtc);
> > @@ -6010,7 +6010,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state,
> >   static void intel_crtc_disable_planes(struct intel_atomic_state *state,
> >   				      struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct intel_crtc_state *new_crtc_state =
> >   		intel_atomic_get_new_crtc_state(state, crtc);
> >   	unsigned int update_mask = new_crtc_state->update_planes;
> > @@ -6032,7 +6032,7 @@ static void intel_crtc_disable_planes(struct intel_atomic_state *state,
> >   			fb_bits |= plane->frontbuffer_bit;
> >   	}
> > -	intel_frontbuffer_flip(dev_priv, fb_bits);
> > +	intel_frontbuffer_flip(i915, fb_bits);
> >   }
> >   static void intel_encoders_pre_pll_enable(struct drm_crtc *crtc,
> > @@ -6190,7 +6190,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   {
> >   	struct drm_crtc *crtc = pipe_config->base.crtc;
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> >   	struct intel_atomic_state *old_intel_state =
> > @@ -6209,8 +6209,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	 *
> >   	 * Spurious PCH underruns also occur during PCH enabling.
> >   	 */
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> > +	intel_set_pch_fifo_underrun_reporting(i915, pipe, false);
> >   	if (pipe_config->has_pch_encoder)
> >   		intel_prepare_shared_dpll(pipe_config);
> > @@ -6238,8 +6238,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   		 * enabling. */
> >   		ironlake_fdi_pll_enable(pipe_config);
> >   	} else {
> > -		assert_fdi_tx_disabled(dev_priv, pipe);
> > -		assert_fdi_rx_disabled(dev_priv, pipe);
> > +		assert_fdi_tx_disabled(i915, pipe);
> > +		assert_fdi_rx_disabled(i915, pipe);
> >   	}
> >   	ironlake_pfit_enable(pipe_config);
> > @@ -6253,8 +6253,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	/* update DSPCNTR to configure gamma for pipe bottom color */
> >   	intel_disable_primary_plane(pipe_config);
> > -	if (dev_priv->display.initial_watermarks != NULL)
> > -		dev_priv->display.initial_watermarks(old_intel_state, pipe_config);
> > +	if (i915->display.initial_watermarks != NULL)
> > +		i915->display.initial_watermarks(old_intel_state, pipe_config);
> >   	intel_enable_pipe(pipe_config);
> >   	if (pipe_config->has_pch_encoder)
> > @@ -6265,7 +6265,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_encoders_enable(crtc, pipe_config, old_state);
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		cpt_verify_modeset(dev, intel_crtc->pipe);
> >   	/*
> > @@ -6275,11 +6275,11 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	 * in case there are more corner cases we don't know about.
> >   	 */
> >   	if (pipe_config->has_pch_encoder) {
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> >   	}
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> > +	intel_set_pch_fifo_underrun_reporting(i915, pipe, true);
> >   }
> >   /* IPS only exists on ULT machines and is tied to pipe A. */
> > @@ -6288,7 +6288,7 @@ static bool hsw_crtc_supports_ips(struct intel_crtc *crtc)
> >   	return HAS_IPS(to_i915(crtc->base.dev)) && crtc->pipe == PIPE_A;
> >   }
> > -static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv,
> > +static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *i915,
> >   					    enum pipe pipe, bool apply)
> >   {
> >   	u32 val = I915_READ(CLKGATE_DIS_PSL(pipe));
> > @@ -6304,7 +6304,7 @@ static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv,
> >   static void icl_pipe_mbus_enable(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > @@ -6319,7 +6319,7 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> >   				struct drm_atomic_state *old_state)
> >   {
> >   	struct drm_crtc *crtc = pipe_config->base.crtc;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe, hsw_workaround_pipe;
> >   	enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> > @@ -6359,18 +6359,18 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	if (!transcoder_is_dsi(cpu_transcoder))
> >   		haswell_set_pipeconf(pipe_config);
> > -	if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv))
> > +	if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915))
> >   		bdw_set_pipemisc(pipe_config);
> >   	intel_crtc->active = true;
> >   	/* Display WA #1180: WaDisableScalarClockGating: glk, cnl */
> > -	psl_clkgate_wa = (IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) &&
> > +	psl_clkgate_wa = (IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) &&
> >   			 pipe_config->pch_pfit.enabled;
> >   	if (psl_clkgate_wa)
> > -		glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, true);
> > +		glk_pipe_scaler_clock_gating_wa(i915, pipe, true);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		skylake_pfit_enable(pipe_config);
> >   	else
> >   		ironlake_pfit_enable(pipe_config);
> > @@ -6382,20 +6382,20 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_color_load_luts(pipe_config);
> >   	intel_color_commit(pipe_config);
> >   	/* update DSPCNTR to configure gamma/csc for pipe bottom color */
> > -	if (INTEL_GEN(dev_priv) < 9)
> > +	if (INTEL_GEN(i915) < 9)
> >   		intel_disable_primary_plane(pipe_config);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_set_pipe_chicken(intel_crtc);
> >   	intel_ddi_set_pipe_settings(pipe_config);
> >   	if (!transcoder_is_dsi(cpu_transcoder))
> >   		intel_ddi_enable_transcoder_func(pipe_config);
> > -	if (dev_priv->display.initial_watermarks != NULL)
> > -		dev_priv->display.initial_watermarks(old_intel_state, pipe_config);
> > +	if (i915->display.initial_watermarks != NULL)
> > +		i915->display.initial_watermarks(old_intel_state, pipe_config);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_pipe_mbus_enable(intel_crtc);
> >   	/* XXX: Do the pipe assertions at the right place for BXT DSI. */
> > @@ -6414,23 +6414,23 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_encoders_enable(crtc, pipe_config, old_state);
> >   	if (psl_clkgate_wa) {
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > -		glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, false);
> > +		intel_wait_for_vblank(i915, pipe);
> > +		glk_pipe_scaler_clock_gating_wa(i915, pipe, false);
> >   	}
> >   	/* If we change the relative order between pipe/planes enabling, we need
> >   	 * to change the workaround. */
> >   	hsw_workaround_pipe = pipe_config->hsw_workaround_pipe;
> > -	if (IS_HASWELL(dev_priv) && hsw_workaround_pipe != INVALID_PIPE) {
> > -		intel_wait_for_vblank(dev_priv, hsw_workaround_pipe);
> > -		intel_wait_for_vblank(dev_priv, hsw_workaround_pipe);
> > +	if (IS_HASWELL(i915) && hsw_workaround_pipe != INVALID_PIPE) {
> > +		intel_wait_for_vblank(i915, hsw_workaround_pipe);
> > +		intel_wait_for_vblank(i915, hsw_workaround_pipe);
> >   	}
> >   }
> >   static void ironlake_pfit_disable(const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* To avoid upsetting the power well on haswell only disable the pfit if
> > @@ -6447,7 +6447,7 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   {
> >   	struct drm_crtc *crtc = old_crtc_state->base.crtc;
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> > @@ -6456,8 +6456,8 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   	 * pipe is already disabled, but FDI RX/TX is still enabled.
> >   	 * Happens at least with VGA+HDMI cloning. Suppress them.
> >   	 */
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> > +	intel_set_pch_fifo_underrun_reporting(i915, pipe, false);
> >   	intel_encoders_disable(crtc, old_crtc_state, old_state);
> > @@ -6474,9 +6474,9 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   	intel_encoders_post_disable(crtc, old_crtc_state, old_state);
> >   	if (old_crtc_state->has_pch_encoder) {
> > -		ironlake_disable_pch_transcoder(dev_priv, pipe);
> > +		ironlake_disable_pch_transcoder(i915, pipe);
> > -		if (HAS_PCH_CPT(dev_priv)) {
> > +		if (HAS_PCH_CPT(i915)) {
> >   			i915_reg_t reg;
> >   			u32 temp;
> > @@ -6497,15 +6497,15 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   		ironlake_fdi_pll_disable(intel_crtc);
> >   	}
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > -	intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> > +	intel_set_pch_fifo_underrun_reporting(i915, pipe, true);
> >   }
> >   static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   				 struct drm_atomic_state *old_state)
> >   {
> >   	struct drm_crtc *crtc = old_crtc_state->base.crtc;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder;
> > @@ -6526,7 +6526,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   	intel_dsc_disable(old_crtc_state);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		skylake_scaler_disable(intel_crtc);
> >   	else
> >   		ironlake_pfit_disable(old_crtc_state);
> > @@ -6539,7 +6539,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	if (!crtc_state->gmch_pfit.control)
> >   		return;
> > @@ -6549,7 +6549,7 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   	 * according to register description and PRM.
> >   	 */
> >   	WARN_ON(I915_READ(PFIT_CONTROL) & PFIT_ENABLE);
> > -	assert_pipe_disabled(dev_priv, crtc->pipe);
> > +	assert_pipe_disabled(i915, crtc->pipe);
> >   	I915_WRITE(PFIT_PGM_RATIOS, crtc_state->gmch_pfit.pgm_ratios);
> >   	I915_WRITE(PFIT_CONTROL, crtc_state->gmch_pfit.control);
> > @@ -6559,31 +6559,31 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state)
> >   	I915_WRITE(BCLRPAT(crtc->pipe), 0);
> >   }
> > -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port)
> > +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port)
> >   {
> >   	if (port == PORT_NONE)
> >   		return false;
> > -	if (IS_ELKHARTLAKE(dev_priv))
> > +	if (IS_ELKHARTLAKE(i915))
> >   		return port <= PORT_C;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return port <= PORT_B;
> >   	return false;
> >   }
> > -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port)
> > +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 11 && !IS_ELKHARTLAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 11 && !IS_ELKHARTLAKE(i915))
> >   		return port >= PORT_C && port <= PORT_F;
> >   	return false;
> >   }
> > -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv, enum port port)
> > +enum tc_port intel_port_to_tc(struct drm_i915_private *i915, enum port port)
> >   {
> > -	if (!intel_port_is_tc(dev_priv, port))
> > +	if (!intel_port_is_tc(i915, port))
> >   		return PORT_TC_NONE;
> >   	return port - PORT_C;
> > @@ -6636,7 +6636,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc,
> >   				  struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_encoder *encoder;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	enum pipe pipe = intel_crtc->pipe;
> > @@ -6658,7 +6658,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc,
> >   		mask |= BIT_ULL(intel_encoder->power_domain);
> >   	}
> > -	if (HAS_DDI(dev_priv) && crtc_state->has_audio)
> > +	if (HAS_DDI(i915) && crtc_state->has_audio)
> >   		mask |= BIT_ULL(POWER_DOMAIN_AUDIO);
> >   	if (crtc_state->shared_dpll)
> > @@ -6671,7 +6671,7 @@ static u64
> >   modeset_get_crtc_power_domains(struct drm_crtc *crtc,
> >   			       struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	enum intel_display_power_domain domain;
> >   	u64 domains, new_domains, old_domains;
> > @@ -6683,18 +6683,18 @@ modeset_get_crtc_power_domains(struct drm_crtc *crtc,
> >   	domains = new_domains & ~old_domains;
> >   	for_each_power_domain(domain, domains)
> > -		intel_display_power_get(dev_priv, domain);
> > +		intel_display_power_get(i915, domain);
> >   	return old_domains & ~new_domains;
> >   }
> > -static void modeset_put_power_domains(struct drm_i915_private *dev_priv,
> > +static void modeset_put_power_domains(struct drm_i915_private *i915,
> >   				      u64 domains)
> >   {
> >   	enum intel_display_power_domain domain;
> >   	for_each_power_domain(domain, domains)
> > -		intel_display_power_put_unchecked(dev_priv, domain);
> > +		intel_display_power_put_unchecked(i915, domain);
> >   }
> >   static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> > @@ -6704,7 +6704,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> >   		to_intel_atomic_state(old_state);
> >   	struct drm_crtc *crtc = pipe_config->base.crtc;
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> > @@ -6717,7 +6717,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_set_pipe_timings(pipe_config);
> >   	intel_set_pipe_src_size(pipe_config);
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> >   		I915_WRITE(CHV_BLEND(pipe), CHV_BLEND_LEGACY);
> >   		I915_WRITE(CHV_CANVAS(pipe), 0);
> >   	}
> > @@ -6726,11 +6726,11 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_crtc->active = true;
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> >   	intel_encoders_pre_pll_enable(crtc, pipe_config, old_state);
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		chv_prepare_pll(intel_crtc, pipe_config);
> >   		chv_enable_pll(intel_crtc, pipe_config);
> >   	} else {
> > @@ -6747,7 +6747,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	/* update DSPCNTR to configure gamma for pipe bottom color */
> >   	intel_disable_primary_plane(pipe_config);
> > -	dev_priv->display.initial_watermarks(old_intel_state,
> > +	i915->display.initial_watermarks(old_intel_state,
> >   					     pipe_config);
> >   	intel_enable_pipe(pipe_config);
> > @@ -6760,7 +6760,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config,
> >   static void i9xx_set_pll_dividers(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	I915_WRITE(FP0(crtc->pipe), crtc_state->dpll_hw_state.fp0);
> >   	I915_WRITE(FP1(crtc->pipe), crtc_state->dpll_hw_state.fp1);
> > @@ -6773,7 +6773,7 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> >   		to_intel_atomic_state(old_state);
> >   	struct drm_crtc *crtc = pipe_config->base.crtc;
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	enum pipe pipe = intel_crtc->pipe;
> > @@ -6792,8 +6792,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	intel_crtc->active = true;
> > -	if (!IS_GEN(dev_priv, 2))
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	if (!IS_GEN(i915, 2))
> > +		intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> >   	intel_encoders_pre_enable(crtc, pipe_config, old_state);
> > @@ -6806,8 +6806,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> >   	/* update DSPCNTR to configure gamma for pipe bottom color */
> >   	intel_disable_primary_plane(pipe_config);
> > -	if (dev_priv->display.initial_watermarks != NULL)
> > -		dev_priv->display.initial_watermarks(old_intel_state,
> > +	if (i915->display.initial_watermarks != NULL)
> > +		i915->display.initial_watermarks(old_intel_state,
> >   						     pipe_config);
> >   	else
> >   		intel_update_watermarks(intel_crtc);
> > @@ -6822,12 +6822,12 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config,
> >   static void i9xx_pfit_disable(const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	if (!old_crtc_state->gmch_pfit.control)
> >   		return;
> > -	assert_pipe_disabled(dev_priv, crtc->pipe);
> > +	assert_pipe_disabled(i915, crtc->pipe);
> >   	DRM_DEBUG_KMS("disabling pfit, current: 0x%08x\n",
> >   		      I915_READ(PFIT_CONTROL));
> > @@ -6839,7 +6839,7 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   {
> >   	struct drm_crtc *crtc = old_crtc_state->base.crtc;
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> > @@ -6847,8 +6847,8 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   	 * On gen2 planes are double buffered but the pipe isn't, so we must
> >   	 * wait for planes to fully turn off before disabling the pipe.
> >   	 */
> > -	if (IS_GEN(dev_priv, 2))
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +	if (IS_GEN(i915, 2))
> > +		intel_wait_for_vblank(i915, pipe);
> >   	intel_encoders_disable(crtc, old_crtc_state, old_state);
> > @@ -6862,25 +6862,25 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state,
> >   	intel_encoders_post_disable(crtc, old_crtc_state, old_state);
> >   	if (!intel_crtc_has_type(old_crtc_state, INTEL_OUTPUT_DSI)) {
> > -		if (IS_CHERRYVIEW(dev_priv))
> > -			chv_disable_pll(dev_priv, pipe);
> > -		else if (IS_VALLEYVIEW(dev_priv))
> > -			vlv_disable_pll(dev_priv, pipe);
> > +		if (IS_CHERRYVIEW(i915))
> > +			chv_disable_pll(i915, pipe);
> > +		else if (IS_VALLEYVIEW(i915))
> > +			vlv_disable_pll(i915, pipe);
> >   		else
> >   			i9xx_disable_pll(old_crtc_state);
> >   	}
> >   	intel_encoders_post_pll_disable(crtc, old_crtc_state, old_state);
> > -	if (!IS_GEN(dev_priv, 2))
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false);
> > +	if (!IS_GEN(i915, 2))
> > +		intel_set_cpu_fifo_underrun_reporting(i915, pipe, false);
> > -	if (!dev_priv->display.initial_watermarks)
> > +	if (!i915->display.initial_watermarks)
> >   		intel_update_watermarks(intel_crtc);
> >   	/* clock the pipe down to 640x480 at 60 to potentially save power */
> > -	if (IS_I830(dev_priv))
> > -		i830_enable_pipe(dev_priv, pipe);
> > +	if (IS_I830(i915))
> > +		i830_enable_pipe(i915, pipe);
> >   }
> >   static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> > @@ -6888,9 +6888,9 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> >   {
> >   	struct intel_encoder *encoder;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_bw_state *bw_state =
> > -		to_intel_bw_state(dev_priv->bw_obj.state);
> > +		to_intel_bw_state(i915->bw_obj.state);
> >   	enum intel_display_power_domain domain;
> >   	struct intel_plane *plane;
> >   	u64 domains;
> > @@ -6901,7 +6901,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> >   	if (!intel_crtc->active)
> >   		return;
> > -	for_each_intel_plane_on_crtc(&dev_priv->drm, intel_crtc, plane) {
> > +	for_each_intel_plane_on_crtc(&i915->drm, intel_crtc, plane) {
> >   		const struct intel_plane_state *plane_state =
> >   			to_intel_plane_state(plane->base.state);
> > @@ -6924,7 +6924,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> >   	WARN_ON(IS_ERR(crtc_state) || ret);
> > -	dev_priv->display.crtc_disable(crtc_state, state);
> > +	i915->display.crtc_disable(crtc_state, state);
> >   	drm_atomic_state_put(state);
> > @@ -6947,12 +6947,12 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> >   	domains = intel_crtc->enabled_power_domains;
> >   	for_each_power_domain(domain, domains)
> > -		intel_display_power_put_unchecked(dev_priv, domain);
> > +		intel_display_power_put_unchecked(i915, domain);
> >   	intel_crtc->enabled_power_domains = 0;
> > -	dev_priv->active_crtcs &= ~(1 << intel_crtc->pipe);
> > -	dev_priv->min_cdclk[intel_crtc->pipe] = 0;
> > -	dev_priv->min_voltage_level[intel_crtc->pipe] = 0;
> > +	i915->active_crtcs &= ~(1 << intel_crtc->pipe);
> > +	i915->min_cdclk[intel_crtc->pipe] = 0;
> > +	i915->min_voltage_level[intel_crtc->pipe] = 0;
> >   	bw_state->data_rate[intel_crtc->pipe] = 0;
> >   	bw_state->num_active_planes[intel_crtc->pipe] = 0;
> > @@ -6964,7 +6964,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc,
> >    */
> >   int intel_display_suspend(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_atomic_state *state;
> >   	int ret;
> > @@ -6973,7 +6973,7 @@ int intel_display_suspend(struct drm_device *dev)
> >   	if (ret)
> >   		DRM_ERROR("Suspending crtc's failed with %i\n", ret);
> >   	else
> > -		dev_priv->modeset_restore_state = state;
> > +		i915->modeset_restore_state = state;
> >   	return ret;
> >   }
> > @@ -7035,7 +7035,7 @@ static int pipe_required_fdi_lanes(struct intel_crtc_state *crtc_state)
> >   static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> >   				     struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_atomic_state *state = pipe_config->base.state;
> >   	struct intel_crtc *other_crtc;
> >   	struct intel_crtc_state *other_crtc_state;
> > @@ -7048,7 +7048,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> >   		return -EINVAL;
> >   	}
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		if (pipe_config->fdi_lanes > 2) {
> >   			DRM_DEBUG_KMS("only 2 lanes on haswell, required: %i lanes\n",
> >   				      pipe_config->fdi_lanes);
> > @@ -7058,7 +7058,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> >   		}
> >   	}
> > -	if (INTEL_INFO(dev_priv)->num_pipes == 2)
> > +	if (INTEL_INFO(i915)->num_pipes == 2)
> >   		return 0;
> >   	/* Ivybridge 3 pipe is really complicated */
> > @@ -7069,7 +7069,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> >   		if (pipe_config->fdi_lanes <= 2)
> >   			return 0;
> > -		other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_C);
> > +		other_crtc = intel_get_crtc_for_pipe(i915, PIPE_C);
> >   		other_crtc_state =
> >   			intel_atomic_get_crtc_state(state, other_crtc);
> >   		if (IS_ERR(other_crtc_state))
> > @@ -7088,7 +7088,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
> >   			return -EINVAL;
> >   		}
> > -		other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_B);
> > +		other_crtc = intel_get_crtc_for_pipe(i915, PIPE_B);
> >   		other_crtc_state =
> >   			intel_atomic_get_crtc_state(state, other_crtc);
> >   		if (IS_ERR(other_crtc_state))
> > @@ -7156,7 +7156,7 @@ static int ironlake_fdi_compute_config(struct intel_crtc *intel_crtc,
> >   bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	/* IPS only exists on ULT machines and is tied to pipe A. */
> >   	if (!hsw_crtc_supports_ips(crtc))
> > @@ -7175,8 +7175,8 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
> >   	 *
> >   	 * Should measure whether using a lower cdclk w/o IPS
> >   	 */
> > -	if (IS_BROADWELL(dev_priv) &&
> > -	    crtc_state->pixel_rate > dev_priv->max_cdclk_freq * 95 / 100)
> > +	if (IS_BROADWELL(i915) &&
> > +	    crtc_state->pixel_rate > i915->max_cdclk_freq * 95 / 100)
> >   		return false;
> >   	return true;
> > @@ -7184,7 +7184,7 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state)
> >   static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_atomic_state *intel_state =
> >   		to_intel_atomic_state(crtc_state->base.state);
> > @@ -7206,7 +7206,7 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
> >   		return false;
> >   	/* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */
> > -	if (IS_BROADWELL(dev_priv) &&
> > +	if (IS_BROADWELL(i915) &&
> >   	    crtc_state->pixel_rate > intel_state->cdclk.logical.cdclk * 95 / 100)
> >   		return false;
> > @@ -7215,11 +7215,11 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state)
> >   static bool intel_crtc_supports_double_wide(const struct intel_crtc *crtc)
> >   {
> > -	const struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	const struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	/* GDG double wide on either pipe, otherwise pipe A only */
> > -	return INTEL_GEN(dev_priv) < 4 &&
> > -		(crtc->pipe == PIPE_A || IS_I915G(dev_priv));
> > +	return INTEL_GEN(i915) < 4 &&
> > +		(crtc->pipe == PIPE_A || IS_I915G(i915));
> >   }
> >   static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config)
> > @@ -7259,9 +7259,9 @@ static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config)
> >   static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		/* FIXME calculate proper pipe pixel rate for GMCH pfit */
> >   		crtc_state->pixel_rate =
> >   			crtc_state->base.adjusted_mode.crtc_clock;
> > @@ -7273,12 +7273,12 @@ static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state)
> >   static int intel_crtc_compute_config(struct intel_crtc *crtc,
> >   				     struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> > -	int clock_limit = dev_priv->max_dotclk_freq;
> > +	int clock_limit = i915->max_dotclk_freq;
> > -	if (INTEL_GEN(dev_priv) < 4) {
> > -		clock_limit = dev_priv->max_cdclk_freq * 9 / 10;
> > +	if (INTEL_GEN(i915) < 4) {
> > +		clock_limit = i915->max_cdclk_freq * 9 / 10;
> >   		/*
> >   		 * Enable double wide mode when the dot clock
> > @@ -7286,7 +7286,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> >   		 */
> >   		if (intel_crtc_supports_double_wide(crtc) &&
> >   		    adjusted_mode->crtc_clock > clock_limit) {
> > -			clock_limit = dev_priv->max_dotclk_freq;
> > +			clock_limit = i915->max_dotclk_freq;
> >   			pipe_config->double_wide = true;
> >   		}
> >   	}
> > @@ -7323,7 +7323,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> >   		}
> >   		if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_LVDS) &&
> > -		    intel_is_dual_link_lvds(dev_priv)) {
> > +		    intel_is_dual_link_lvds(i915)) {
> >   			DRM_DEBUG_KMS("Odd pipe source width not supported with dual link LVDS\n");
> >   			return -EINVAL;
> >   		}
> > @@ -7332,7 +7332,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc,
> >   	/* Cantiga+ cannot handle modes with a hsync front porch of 0.
> >   	 * WaPruneModeWithIncorrectHsyncOffset:ctg,elk,ilk,snb,ivb,vlv,hsw.
> >   	 */
> > -	if ((INTEL_GEN(dev_priv) > 4 || IS_G4X(dev_priv)) &&
> > +	if ((INTEL_GEN(i915) > 4 || IS_G4X(i915)) &&
> >   		adjusted_mode->crtc_hsync_start == adjusted_mode->crtc_hdisplay)
> >   		return -EINVAL;
> > @@ -7392,12 +7392,12 @@ intel_link_compute_m_n(u16 bits_per_pixel, int nlanes,
> >   		    constant_n);
> >   }
> > -static inline bool intel_panel_use_ssc(struct drm_i915_private *dev_priv)
> > +static inline bool intel_panel_use_ssc(struct drm_i915_private *i915)
> >   {
> >   	if (i915_modparams.panel_use_ssc >= 0)
> >   		return i915_modparams.panel_use_ssc != 0;
> > -	return dev_priv->vbt.lvds_use_ssc
> > -		&& !(dev_priv->quirks & QUIRK_LVDS_SSC_DISABLE);
> > +	return i915->vbt.lvds_use_ssc
> > +		&& !(i915->quirks & QUIRK_LVDS_SSC_DISABLE);
> >   }
> >   static u32 pnv_dpll_compute_fp(struct dpll *dpll)
> > @@ -7414,10 +7414,10 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc,
> >   				     struct intel_crtc_state *crtc_state,
> >   				     struct dpll *reduced_clock)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 fp, fp2 = 0;
> > -	if (IS_PINEVIEW(dev_priv)) {
> > +	if (IS_PINEVIEW(i915)) {
> >   		fp = pnv_dpll_compute_fp(&crtc_state->dpll);
> >   		if (reduced_clock)
> >   			fp2 = pnv_dpll_compute_fp(reduced_clock);
> > @@ -7437,7 +7437,7 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc,
> >   	}
> >   }
> > -static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe
> > +static void vlv_pllb_recal_opamp(struct drm_i915_private *i915, enum pipe
> >   		pipe)
> >   {
> >   	u32 reg_val;
> > @@ -7446,31 +7446,31 @@ static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe
> >   	 * PLLB opamp always calibrates to max value of 0x3f, force enable it
> >   	 * and set it to a reasonable value instead.
> >   	 */
> > -	reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1));
> > +	reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1));
> >   	reg_val &= 0xffffff00;
> >   	reg_val |= 0x00000030;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val);
> > -	reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13);
> > +	reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13);
> >   	reg_val &= 0x00ffffff;
> >   	reg_val |= 0x8c000000;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val);
> > +	vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val);
> > -	reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1));
> > +	reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1));
> >   	reg_val &= 0xffffff00;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val);
> > -	reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13);
> > +	reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13);
> >   	reg_val &= 0x00ffffff;
> >   	reg_val |= 0xb0000000;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val);
> > +	vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val);
> >   }
> >   static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_state,
> >   					 const struct intel_link_m_n *m_n)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	I915_WRITE(PCH_TRANS_DATA_M1(pipe), TU_SIZE(m_n->tu) | m_n->gmch_m);
> > @@ -7479,17 +7479,17 @@ static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> >   	I915_WRITE(PCH_TRANS_LINK_N1(pipe), m_n->link_n);
> >   }
> > -static bool transcoder_has_m2_n2(struct drm_i915_private *dev_priv,
> > +static bool transcoder_has_m2_n2(struct drm_i915_private *i915,
> >   				 enum transcoder transcoder)
> >   {
> > -	if (IS_HASWELL(dev_priv))
> > +	if (IS_HASWELL(i915))
> >   		return transcoder == TRANSCODER_EDP;
> >   	/*
> >   	 * Strictly speaking some registers are available before
> >   	 * gen7, but we only support DRRS on gen7+
> >   	 */
> > -	return IS_GEN(dev_priv, 7) || IS_CHERRYVIEW(dev_priv);
> > +	return IS_GEN(i915, 7) || IS_CHERRYVIEW(i915);
> >   }
> >   static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_state,
> > @@ -7497,11 +7497,11 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> >   					 const struct intel_link_m_n *m2_n2)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum transcoder transcoder = crtc_state->cpu_transcoder;
> > -	if (INTEL_GEN(dev_priv) >= 5) {
> > +	if (INTEL_GEN(i915) >= 5) {
> >   		I915_WRITE(PIPE_DATA_M1(transcoder), TU_SIZE(m_n->tu) | m_n->gmch_m);
> >   		I915_WRITE(PIPE_DATA_N1(transcoder), m_n->gmch_n);
> >   		I915_WRITE(PIPE_LINK_M1(transcoder), m_n->link_m);
> > @@ -7511,7 +7511,7 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta
> >   		 * (to make sure the registers are not unnecessarily accessed).
> >   		 */
> >   		if (m2_n2 && crtc_state->has_drrs &&
> > -		    transcoder_has_m2_n2(dev_priv, transcoder)) {
> > +		    transcoder_has_m2_n2(i915, transcoder)) {
> >   			I915_WRITE(PIPE_DATA_M2(transcoder),
> >   					TU_SIZE(m2_n2->tu) | m2_n2->gmch_m);
> >   			I915_WRITE(PIPE_DATA_N2(transcoder), m2_n2->gmch_n);
> > @@ -7588,7 +7588,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> >   			    const struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 mdiv;
> >   	u32 bestn, bestm1, bestm2, bestp1, bestp2;
> > @@ -7603,7 +7603,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> >   	if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> >   		return;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	bestn = pipe_config->dpll.n;
> >   	bestm1 = pipe_config->dpll.m1;
> > @@ -7615,18 +7615,18 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> >   	/* PLL B needs special handling */
> >   	if (pipe == PIPE_B)
> > -		vlv_pllb_recal_opamp(dev_priv, pipe);
> > +		vlv_pllb_recal_opamp(i915, pipe);
> >   	/* Set up Tx target for periodic Rcomp update */
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9_BCAST, 0x0100000f);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW9_BCAST, 0x0100000f);
> >   	/* Disable target IRef on PLL */
> > -	reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW8(pipe));
> > +	reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW8(pipe));
> >   	reg_val &= 0x00ffffff;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW8(pipe), reg_val);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW8(pipe), reg_val);
> >   	/* Disable fast lock */
> > -	vlv_dpio_write(dev_priv, pipe, VLV_CMN_DW0, 0x610);
> > +	vlv_dpio_write(i915, pipe, VLV_CMN_DW0, 0x610);
> >   	/* Set idtafcrecal before PLL is enabled */
> >   	mdiv = ((bestm1 << DPIO_M1DIV_SHIFT) | (bestm2 & DPIO_M2DIV_MASK));
> > @@ -7640,55 +7640,55 @@ static void vlv_prepare_pll(struct intel_crtc *crtc,
> >   	 * Note: don't use the DAC post divider as it seems unstable.
> >   	 */
> >   	mdiv |= (DPIO_POST_DIV_HDMIDP << DPIO_POST_DIV_SHIFT);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv);
> >   	mdiv |= DPIO_ENABLE_CALIBRATION;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv);
> >   	/* Set HBR and RBR LPF coefficients */
> >   	if (pipe_config->port_clock == 162000 ||
> >   	    intel_crtc_has_type(pipe_config, INTEL_OUTPUT_ANALOG) ||
> >   	    intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI))
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe),
> > +		vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe),
> >   				 0x009f0003);
> >   	else
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe),
> > +		vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe),
> >   				 0x00d0000f);
> >   	if (intel_crtc_has_dp_encoder(pipe_config)) {
> >   		/* Use SSC source */
> >   		if (pipe == PIPE_A)
> > -			vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> > +			vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> >   					 0x0df40000);
> >   		else
> > -			vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> > +			vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> >   					 0x0df70000);
> >   	} else { /* HDMI or VGA */
> >   		/* Use bend source */
> >   		if (pipe == PIPE_A)
> > -			vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> > +			vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> >   					 0x0df70000);
> >   		else
> > -			vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe),
> > +			vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe),
> >   					 0x0df40000);
> >   	}
> > -	coreclk = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW7(pipe));
> > +	coreclk = vlv_dpio_read(i915, pipe, VLV_PLL_DW7(pipe));
> >   	coreclk = (coreclk & 0x0000ff00) | 0x01c00000;
> >   	if (intel_crtc_has_dp_encoder(pipe_config))
> >   		coreclk |= 0x01000000;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW7(pipe), coreclk);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW7(pipe), coreclk);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW11(pipe), 0x87871000);
> > +	vlv_dpio_write(i915, pipe, VLV_PLL_DW11(pipe), 0x87871000);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   static void chv_prepare_pll(struct intel_crtc *crtc,
> >   			    const struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum dpio_channel port = vlv_pipe_to_channel(pipe);
> >   	u32 loopfilter, tribuf_calcntr;
> > @@ -7714,42 +7714,42 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> >   	dpio_val = 0;
> >   	loopfilter = 0;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* p1 and p2 divider */
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW13(port),
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW13(port),
> >   			5 << DPIO_CHV_S1_DIV_SHIFT |
> >   			bestp1 << DPIO_CHV_P1_DIV_SHIFT |
> >   			bestp2 << DPIO_CHV_P2_DIV_SHIFT |
> >   			1 << DPIO_CHV_K_DIV_SHIFT);
> >   	/* Feedback post-divider - m2 */
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW0(port), bestm2);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW0(port), bestm2);
> >   	/* Feedback refclk divider - n and m1 */
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW1(port),
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW1(port),
> >   			DPIO_CHV_M1_DIV_BY_2 |
> >   			1 << DPIO_CHV_N_DIV_SHIFT);
> >   	/* M2 fraction division */
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW2(port), bestm2_frac);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW2(port), bestm2_frac);
> >   	/* M2 fraction division enable */
> > -	dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port));
> > +	dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port));
> >   	dpio_val &= ~(DPIO_CHV_FEEDFWD_GAIN_MASK | DPIO_CHV_FRAC_DIV_EN);
> >   	dpio_val |= (2 << DPIO_CHV_FEEDFWD_GAIN_SHIFT);
> >   	if (bestm2_frac)
> >   		dpio_val |= DPIO_CHV_FRAC_DIV_EN;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW3(port), dpio_val);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW3(port), dpio_val);
> >   	/* Program digital lock detect threshold */
> > -	dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW9(port));
> > +	dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW9(port));
> >   	dpio_val &= ~(DPIO_CHV_INT_LOCK_THRESHOLD_MASK |
> >   					DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE);
> >   	dpio_val |= (0x5 << DPIO_CHV_INT_LOCK_THRESHOLD_SHIFT);
> >   	if (!bestm2_frac)
> >   		dpio_val |= DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW9(port), dpio_val);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW9(port), dpio_val);
> >   	/* Loop filter */
> >   	if (vco == 5400000) {
> > @@ -7774,24 +7774,24 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> >   		loopfilter |= (0x3 << DPIO_CHV_GAIN_CTRL_SHIFT);
> >   		tribuf_calcntr = 0;
> >   	}
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW6(port), loopfilter);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW6(port), loopfilter);
> > -	dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW8(port));
> > +	dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW8(port));
> >   	dpio_val &= ~DPIO_CHV_TDC_TARGET_CNT_MASK;
> >   	dpio_val |= (tribuf_calcntr << DPIO_CHV_TDC_TARGET_CNT_SHIFT);
> > -	vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW8(port), dpio_val);
> > +	vlv_dpio_write(i915, pipe, CHV_PLL_DW8(port), dpio_val);
> >   	/* AFC Recal */
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port),
> > -			vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port)) |
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port),
> > +			vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port)) |
> >   			DPIO_AFC_RECAL);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   /**
> >    * vlv_force_pll_on - forcibly enable just the PLL
> > - * @dev_priv: i915 private structure
> > + * @i915: i915 private structure
> >    * @pipe: pipe PLL to enable
> >    * @dpll: PLL configuration
> >    *
> > @@ -7799,10 +7799,10 @@ static void chv_prepare_pll(struct intel_crtc *crtc,
> >    * in cases where we need the PLL enabled even when @pipe is not going to
> >    * be enabled.
> >    */
> > -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> > +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe,
> >   		     const struct dpll *dpll)
> >   {
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	struct intel_crtc_state *pipe_config;
> >   	pipe_config = kzalloc(sizeof(*pipe_config), GFP_KERNEL);
> > @@ -7813,7 +7813,7 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> >   	pipe_config->pixel_multiplier = 1;
> >   	pipe_config->dpll = *dpll;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		chv_compute_dpll(crtc, pipe_config);
> >   		chv_prepare_pll(crtc, pipe_config);
> >   		chv_enable_pll(crtc, pipe_config);
> > @@ -7830,25 +7830,25 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> >   /**
> >    * vlv_force_pll_off - forcibly disable just the PLL
> > - * @dev_priv: i915 private structure
> > + * @i915: i915 private structure
> >    * @pipe: pipe PLL to disable
> >    *
> >    * Disable the PLL for @pipe. To be used in cases where we need
> >    * the PLL enabled even when @pipe is not going to be enabled.
> >    */
> > -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> > -	if (IS_CHERRYVIEW(dev_priv))
> > -		chv_disable_pll(dev_priv, pipe);
> > +	if (IS_CHERRYVIEW(i915))
> > +		chv_disable_pll(i915, pipe);
> >   	else
> > -		vlv_disable_pll(dev_priv, pipe);
> > +		vlv_disable_pll(i915, pipe);
> >   }
> >   static void i9xx_compute_dpll(struct intel_crtc *crtc,
> >   			      struct intel_crtc_state *crtc_state,
> >   			      struct dpll *reduced_clock)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 dpll;
> >   	struct dpll *clock = &crtc_state->dpll;
> > @@ -7861,8 +7861,8 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> >   	else
> >   		dpll |= DPLLB_MODE_DAC_SERIAL;
> > -	if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> > -	    IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> > +	if (IS_I945G(i915) || IS_I945GM(i915) ||
> > +	    IS_G33(i915) || IS_PINEVIEW(i915)) {
> >   		dpll |= (crtc_state->pixel_multiplier - 1)
> >   			<< SDVO_MULTIPLIER_SHIFT_HIRES;
> >   	}
> > @@ -7875,11 +7875,11 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> >   		dpll |= DPLL_SDVO_HIGH_SPEED;
> >   	/* compute bitmask from p1 value */
> > -	if (IS_PINEVIEW(dev_priv))
> > +	if (IS_PINEVIEW(i915))
> >   		dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW;
> >   	else {
> >   		dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT;
> > -		if (IS_G4X(dev_priv) && reduced_clock)
> > +		if (IS_G4X(i915) && reduced_clock)
> >   			dpll |= (1 << (reduced_clock->p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT;
> >   	}
> >   	switch (clock->p2) {
> > @@ -7896,13 +7896,13 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> >   		dpll |= DPLLB_LVDS_P2_CLOCK_DIV_14;
> >   		break;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		dpll |= (6 << PLL_LOAD_PULSE_PHASE_SHIFT);
> >   	if (crtc_state->sdvo_tv_clock)
> >   		dpll |= PLL_REF_INPUT_TVCLKINBC;
> >   	else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> > -		 intel_panel_use_ssc(dev_priv))
> > +		 intel_panel_use_ssc(i915))
> >   		dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> >   	else
> >   		dpll |= PLL_REF_INPUT_DREFCLK;
> > @@ -7910,7 +7910,7 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc,
> >   	dpll |= DPLL_VCO_ENABLE;
> >   	crtc_state->dpll_hw_state.dpll = dpll;
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		u32 dpll_md = (crtc_state->pixel_multiplier - 1)
> >   			<< DPLL_MD_UDI_MULTIPLIER_SHIFT;
> >   		crtc_state->dpll_hw_state.dpll_md = dpll_md;
> > @@ -7922,7 +7922,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> >   			      struct dpll *reduced_clock)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 dpll;
> >   	struct dpll *clock = &crtc_state->dpll;
> > @@ -7953,12 +7953,12 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> >   	 * both DPLLS. The spec says we should disable the DVO 2X clock
> >   	 * when not needed, but this seems to work fine in practice.
> >   	 */
> > -	if (IS_I830(dev_priv) ||
> > +	if (IS_I830(i915) ||
> >   	    intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DVO))
> >   		dpll |= DPLL_DVO_2X_MODE;
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> > -	    intel_panel_use_ssc(dev_priv))
> > +	    intel_panel_use_ssc(i915))
> >   		dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> >   	else
> >   		dpll |= PLL_REF_INPUT_DREFCLK;
> > @@ -7970,7 +7970,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc,
> >   static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> > @@ -7996,7 +7996,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> >   			vsyncshift += adjusted_mode->crtc_htotal;
> >   	}
> > -	if (INTEL_GEN(dev_priv) > 3)
> > +	if (INTEL_GEN(i915) > 3)
> >   		I915_WRITE(VSYNCSHIFT(cpu_transcoder), vsyncshift);
> >   	I915_WRITE(HTOTAL(cpu_transcoder),
> > @@ -8023,7 +8023,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> >   	 * programmed with the VTOTAL_EDP value. Same for VTOTAL_C. This is
> >   	 * documented on the DDI_FUNC_CTL register description, EDP Input Select
> >   	 * bits. */
> > -	if (IS_HASWELL(dev_priv) && cpu_transcoder == TRANSCODER_EDP &&
> > +	if (IS_HASWELL(i915) && cpu_transcoder == TRANSCODER_EDP &&
> >   	    (pipe == PIPE_B || pipe == PIPE_C))
> >   		I915_WRITE(VTOTAL(pipe), I915_READ(VTOTAL(cpu_transcoder)));
> > @@ -8032,7 +8032,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state)
> >   static void intel_set_pipe_src_size(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* pipesrc controls the size that is scaled from, which should
> > @@ -8047,7 +8047,7 @@ static void intel_get_pipe_timings(struct intel_crtc *crtc,
> >   				   struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum transcoder cpu_transcoder = pipe_config->cpu_transcoder;
> >   	u32 tmp;
> > @@ -8092,7 +8092,7 @@ static void intel_get_pipe_src_size(struct intel_crtc *crtc,
> >   				    struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 tmp;
> >   	tmp = I915_READ(PIPESRC(crtc->pipe));
> > @@ -8129,21 +8129,21 @@ void intel_mode_from_pipe_config(struct drm_display_mode *mode,
> >   static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 pipeconf;
> >   	pipeconf = 0;
> >   	/* we keep both pipes enabled on 830 */
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		pipeconf |= I915_READ(PIPECONF(crtc->pipe)) & PIPECONF_ENABLE;
> >   	if (crtc_state->double_wide)
> >   		pipeconf |= PIPECONF_DOUBLE_WIDE;
> >   	/* only g4x and later have fancy bpc/dither controls */
> > -	if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> > -	    IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> > +	    IS_CHERRYVIEW(i915)) {
> >   		/* Bspec claims that we can't use dithering for 30bpp pipes. */
> >   		if (crtc_state->dither && crtc_state->pipe_bpp != 30)
> >   			pipeconf |= PIPECONF_DITHER_EN |
> > @@ -8166,7 +8166,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   	}
> >   	if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE) {
> > -		if (INTEL_GEN(dev_priv) < 4 ||
> > +		if (INTEL_GEN(i915) < 4 ||
> >   		    intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO))
> >   			pipeconf |= PIPECONF_INTERLACE_W_FIELD_INDICATION;
> >   		else
> > @@ -8175,7 +8175,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   		pipeconf |= PIPECONF_PROGRESSIVE;
> >   	}
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   	     crtc_state->limited_color_range)
> >   		pipeconf |= PIPECONF_COLOR_RANGE_SELECT;
> > @@ -8189,7 +8189,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	const struct intel_limit *limit;
> >   	int refclk = 48000;
> > @@ -8197,8 +8197,8 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> >   	       sizeof(crtc_state->dpll_hw_state));
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if (intel_panel_use_ssc(dev_priv)) {
> > -			refclk = dev_priv->vbt.lvds_ssc_freq;
> > +		if (intel_panel_use_ssc(i915)) {
> > +			refclk = i915->vbt.lvds_ssc_freq;
> >   			DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> >   		}
> > @@ -8224,7 +8224,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc,
> >   static int g4x_crtc_compute_clock(struct intel_crtc *crtc,
> >   				  struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct intel_limit *limit;
> >   	int refclk = 96000;
> > @@ -8232,12 +8232,12 @@ static int g4x_crtc_compute_clock(struct intel_crtc *crtc,
> >   	       sizeof(crtc_state->dpll_hw_state));
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if (intel_panel_use_ssc(dev_priv)) {
> > -			refclk = dev_priv->vbt.lvds_ssc_freq;
> > +		if (intel_panel_use_ssc(i915)) {
> > +			refclk = i915->vbt.lvds_ssc_freq;
> >   			DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> >   		}
> > -		if (intel_is_dual_link_lvds(dev_priv))
> > +		if (intel_is_dual_link_lvds(i915))
> >   			limit = &intel_limits_g4x_dual_channel_lvds;
> >   		else
> >   			limit = &intel_limits_g4x_single_channel_lvds;
> > @@ -8267,7 +8267,7 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc,
> >   				  struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	const struct intel_limit *limit;
> >   	int refclk = 96000;
> > @@ -8275,8 +8275,8 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc,
> >   	       sizeof(crtc_state->dpll_hw_state));
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if (intel_panel_use_ssc(dev_priv)) {
> > -			refclk = dev_priv->vbt.lvds_ssc_freq;
> > +		if (intel_panel_use_ssc(i915)) {
> > +			refclk = i915->vbt.lvds_ssc_freq;
> >   			DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> >   		}
> > @@ -8301,7 +8301,7 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	const struct intel_limit *limit;
> >   	int refclk = 96000;
> > @@ -8309,8 +8309,8 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc,
> >   	       sizeof(crtc_state->dpll_hw_state));
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if (intel_panel_use_ssc(dev_priv)) {
> > -			refclk = dev_priv->vbt.lvds_ssc_freq;
> > +		if (intel_panel_use_ssc(i915)) {
> > +			refclk = i915->vbt.lvds_ssc_freq;
> >   			DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk);
> >   		}
> > @@ -8373,22 +8373,22 @@ static int vlv_crtc_compute_clock(struct intel_crtc *crtc,
> >   	return 0;
> >   }
> > -static bool i9xx_has_pfit(struct drm_i915_private *dev_priv)
> > +static bool i9xx_has_pfit(struct drm_i915_private *i915)
> >   {
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		return false;
> > -	return INTEL_GEN(dev_priv) >= 4 ||
> > -		IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv);
> > +	return INTEL_GEN(i915) >= 4 ||
> > +		IS_PINEVIEW(i915) || IS_MOBILE(i915);
> >   }
> >   static void i9xx_get_pfit_config(struct intel_crtc *crtc,
> >   				 struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 tmp;
> > -	if (!i9xx_has_pfit(dev_priv))
> > +	if (!i9xx_has_pfit(i915))
> >   		return;
> >   	tmp = I915_READ(PFIT_CONTROL);
> > @@ -8396,7 +8396,7 @@ static void i9xx_get_pfit_config(struct intel_crtc *crtc,
> >   		return;
> >   	/* Check whether the pfit is attached to our pipe. */
> > -	if (INTEL_GEN(dev_priv) < 4) {
> > +	if (INTEL_GEN(i915) < 4) {
> >   		if (crtc->pipe != PIPE_B)
> >   			return;
> >   	} else {
> > @@ -8412,7 +8412,7 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc,
> >   			       struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = pipe_config->cpu_transcoder;
> >   	struct dpll clock;
> >   	u32 mdiv;
> > @@ -8422,9 +8422,9 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc,
> >   	if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> >   		return;
> > -	vlv_dpio_get(dev_priv);
> > -	mdiv = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW3(pipe));
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_get(i915);
> > +	mdiv = vlv_dpio_read(i915, pipe, VLV_PLL_DW3(pipe));
> > +	vlv_dpio_put(i915);
> >   	clock.m1 = (mdiv >> DPIO_M1DIV_SHIFT) & 7;
> >   	clock.m2 = mdiv & DPIO_M2DIV_MASK;
> > @@ -8440,7 +8440,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> >   			      struct intel_initial_plane_config *plane_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> >   	enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> >   	enum pipe pipe;
> > @@ -8467,7 +8467,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> >   	val = I915_READ(DSPCNTR(i9xx_plane));
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		if (val & DISPPLANE_TILED) {
> >   			plane_config->tiling = I915_TILING_X;
> >   			fb->modifier = I915_FORMAT_MOD_X_TILED;
> > @@ -8477,7 +8477,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> >   			plane_config->rotation = DRM_MODE_ROTATE_180;
> >   	}
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B &&
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_B &&
> >   	    val & DISPPLANE_MIRROR)
> >   		plane_config->rotation |= DRM_MODE_REFLECT_X;
> > @@ -8485,10 +8485,10 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc,
> >   	fourcc = i9xx_format_to_fourcc(pixel_format);
> >   	fb->format = drm_format_info(fourcc);
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		offset = I915_READ(DSPOFFSET(i9xx_plane));
> >   		base = I915_READ(DSPSURF(i9xx_plane)) & 0xfffff000;
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > +	} else if (INTEL_GEN(i915) >= 4) {
> >   		if (plane_config->tiling)
> >   			offset = I915_READ(DSPTILEOFF(i9xx_plane));
> >   		else
> > @@ -8522,7 +8522,7 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> >   			       struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = pipe_config->cpu_transcoder;
> >   	enum dpio_channel port = vlv_pipe_to_channel(pipe);
> >   	struct dpll clock;
> > @@ -8533,13 +8533,13 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> >   	if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0)
> >   		return;
> > -	vlv_dpio_get(dev_priv);
> > -	cmn_dw13 = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW13(port));
> > -	pll_dw0 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW0(port));
> > -	pll_dw1 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW1(port));
> > -	pll_dw2 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW2(port));
> > -	pll_dw3 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port));
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_get(i915);
> > +	cmn_dw13 = vlv_dpio_read(i915, pipe, CHV_CMN_DW13(port));
> > +	pll_dw0 = vlv_dpio_read(i915, pipe, CHV_PLL_DW0(port));
> > +	pll_dw1 = vlv_dpio_read(i915, pipe, CHV_PLL_DW1(port));
> > +	pll_dw2 = vlv_dpio_read(i915, pipe, CHV_PLL_DW2(port));
> > +	pll_dw3 = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port));
> > +	vlv_dpio_put(i915);
> >   	clock.m1 = (pll_dw1 & 0x7) == DPIO_CHV_M1_DIV_BY_2 ? 2 : 0;
> >   	clock.m2 = (pll_dw0 & 0xff) << 22;
> > @@ -8555,12 +8555,12 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc,
> >   static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc,
> >   					struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum intel_output_format output = INTEL_OUTPUT_FORMAT_RGB;
> >   	pipe_config->lspcon_downsampling = false;
> > -	if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9) {
> > +	if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9) {
> >   		u32 tmp = I915_READ(PIPEMISC(crtc->pipe));
> >   		if (tmp & PIPEMISC_OUTPUT_COLORSPACE_YUV) {
> > @@ -8571,8 +8571,8 @@ static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc,
> >   				/* We support 4:2:0 in full blend mode only */
> >   				if (!blend)
> >   					output = INTEL_OUTPUT_FORMAT_INVALID;
> > -				else if (!(IS_GEMINILAKE(dev_priv) ||
> > -					   INTEL_GEN(dev_priv) >= 10))
> > +				else if (!(IS_GEMINILAKE(i915) ||
> > +					   INTEL_GEN(i915) >= 10))
> >   					output = INTEL_OUTPUT_FORMAT_INVALID;
> >   				else
> >   					output = INTEL_OUTPUT_FORMAT_YCBCR420;
> > @@ -8599,7 +8599,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum i9xx_plane_id i9xx_plane = plane->i9xx_plane;
> >   	u32 tmp;
> > @@ -8608,7 +8608,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> >   	if (tmp & DISPPLANE_GAMMA_ENABLE)
> >   		crtc_state->gamma_enable = true;
> > -	if (!HAS_GMCH(dev_priv) &&
> > +	if (!HAS_GMCH(i915) &&
> >   	    tmp & DISPPLANE_PIPE_CSC_ENABLE)
> >   		crtc_state->csc_enable = true;
> >   }
> > @@ -8616,14 +8616,14 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state)
> >   static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   				 struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	u32 tmp;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -8637,8 +8637,8 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   	if (!(tmp & PIPECONF_ENABLE))
> >   		goto out;
> > -	if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> > -	    IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> > +	    IS_CHERRYVIEW(i915)) {
> >   		switch (tmp & PIPECONF_BPC_MASK) {
> >   		case PIPECONF_6BPC:
> >   			pipe_config->pipe_bpp = 18;
> > @@ -8654,20 +8654,20 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   		}
> >   	}
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   	    (tmp & PIPECONF_COLOR_RANGE_SELECT))
> >   		pipe_config->limited_color_range = true;
> >   	pipe_config->gamma_mode = (tmp & PIPECONF_GAMMA_MODE_MASK_I9XX) >>
> >   		PIPECONF_GAMMA_MODE_SHIFT;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		pipe_config->cgm_mode = I915_READ(CGM_PIPE_MODE(crtc->pipe));
> >   	i9xx_get_pipe_color_config(pipe_config);
> >   	intel_color_get_config(pipe_config);
> > -	if (INTEL_GEN(dev_priv) < 4)
> > +	if (INTEL_GEN(i915) < 4)
> >   		pipe_config->double_wide = tmp & PIPECONF_DOUBLE_WIDE;
> >   	intel_get_pipe_timings(crtc, pipe_config);
> > @@ -8675,18 +8675,18 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   	i9xx_get_pfit_config(crtc, pipe_config);
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		/* No way to read it out on pipes B and C */
> > -		if (IS_CHERRYVIEW(dev_priv) && crtc->pipe != PIPE_A)
> > -			tmp = dev_priv->chv_dpll_md[crtc->pipe];
> > +		if (IS_CHERRYVIEW(i915) && crtc->pipe != PIPE_A)
> > +			tmp = i915->chv_dpll_md[crtc->pipe];
> >   		else
> >   			tmp = I915_READ(DPLL_MD(crtc->pipe));
> >   		pipe_config->pixel_multiplier =
> >   			((tmp & DPLL_MD_UDI_MULTIPLIER_MASK)
> >   			 >> DPLL_MD_UDI_MULTIPLIER_SHIFT) + 1;
> >   		pipe_config->dpll_hw_state.dpll_md = tmp;
> > -	} else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> > -		   IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> > +	} else if (IS_I945G(i915) || IS_I945GM(i915) ||
> > +		   IS_G33(i915) || IS_PINEVIEW(i915)) {
> >   		tmp = I915_READ(DPLL(crtc->pipe));
> >   		pipe_config->pixel_multiplier =
> >   			((tmp & SDVO_MULTIPLIER_MASK)
> > @@ -8698,7 +8698,7 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   		pipe_config->pixel_multiplier = 1;
> >   	}
> >   	pipe_config->dpll_hw_state.dpll = I915_READ(DPLL(crtc->pipe));
> > -	if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv)) {
> > +	if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915)) {
> >   		pipe_config->dpll_hw_state.fp0 = I915_READ(FP0(crtc->pipe));
> >   		pipe_config->dpll_hw_state.fp1 = I915_READ(FP1(crtc->pipe));
> >   	} else {
> > @@ -8708,9 +8708,9 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   						     DPLL_PORTB_READY_MASK);
> >   	}
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		chv_crtc_clock_get(crtc, pipe_config);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915))
> >   		vlv_crtc_clock_get(crtc, pipe_config);
> >   	else
> >   		i9xx_crtc_clock_get(crtc, pipe_config);
> > @@ -8726,12 +8726,12 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > -static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> > +static void ironlake_init_pch_refclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> >   	int i;
> > @@ -8744,7 +8744,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   	bool using_ssc_source = false;
> >   	/* We need to take the global config into account */
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		switch (encoder->type) {
> >   		case INTEL_OUTPUT_LVDS:
> >   			has_panel = true;
> > @@ -8760,8 +8760,8 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   		}
> >   	}
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > -		has_ck505 = dev_priv->vbt.display_clock_mode;
> > +	if (HAS_PCH_IBX(i915)) {
> > +		has_ck505 = i915->vbt.display_clock_mode;
> >   		can_ssc = has_ck505;
> >   	} else {
> >   		has_ck505 = false;
> > @@ -8769,7 +8769,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   	}
> >   	/* Check if any DPLLs are using the SSC source */
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> >   		u32 temp = I915_READ(PCH_DPLL(i));
> >   		if (!(temp & DPLL_VCO_ENABLE))
> > @@ -8810,11 +8810,11 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   	if (has_panel) {
> >   		final |= DREF_SSC_SOURCE_ENABLE;
> > -		if (intel_panel_use_ssc(dev_priv) && can_ssc)
> > +		if (intel_panel_use_ssc(i915) && can_ssc)
> >   			final |= DREF_SSC1_ENABLE;
> >   		if (has_cpu_edp) {
> > -			if (intel_panel_use_ssc(dev_priv) && can_ssc)
> > +			if (intel_panel_use_ssc(i915) && can_ssc)
> >   				final |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD;
> >   			else
> >   				final |= DREF_CPU_SOURCE_OUTPUT_NONSPREAD;
> > @@ -8841,7 +8841,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   		val |= DREF_SSC_SOURCE_ENABLE;
> >   		/* SSC must be turned on before enabling the CPU output  */
> > -		if (intel_panel_use_ssc(dev_priv) && can_ssc) {
> > +		if (intel_panel_use_ssc(i915) && can_ssc) {
> >   			DRM_DEBUG_KMS("Using SSC on panel\n");
> >   			val |= DREF_SSC1_ENABLE;
> >   		} else
> > @@ -8856,7 +8856,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   		/* Enable CPU source on CPU attached eDP */
> >   		if (has_cpu_edp) {
> > -			if (intel_panel_use_ssc(dev_priv) && can_ssc) {
> > +			if (intel_panel_use_ssc(i915) && can_ssc) {
> >   				DRM_DEBUG_KMS("Using SSC on eDP\n");
> >   				val |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD;
> >   			} else
> > @@ -8898,7 +8898,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   	BUG_ON(val != final);
> >   }
> > -static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv)
> > +static void lpt_reset_fdi_mphy(struct drm_i915_private *i915)
> >   {
> >   	u32 tmp;
> > @@ -8920,78 +8920,78 @@ static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv)
> >   }
> >   /* WaMPhyProgramming:hsw */
> > -static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv)
> > +static void lpt_program_fdi_mphy(struct drm_i915_private *i915)
> >   {
> >   	u32 tmp;
> > -	tmp = intel_sbi_read(dev_priv, 0x8008, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x8008, SBI_MPHY);
> >   	tmp &= ~(0xFF << 24);
> >   	tmp |= (0x12 << 24);
> > -	intel_sbi_write(dev_priv, 0x8008, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x8008, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2008, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2008, SBI_MPHY);
> >   	tmp |= (1 << 11);
> > -	intel_sbi_write(dev_priv, 0x2008, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2008, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2108, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2108, SBI_MPHY);
> >   	tmp |= (1 << 11);
> > -	intel_sbi_write(dev_priv, 0x2108, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2108, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x206C, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x206C, SBI_MPHY);
> >   	tmp |= (1 << 24) | (1 << 21) | (1 << 18);
> > -	intel_sbi_write(dev_priv, 0x206C, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x206C, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x216C, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x216C, SBI_MPHY);
> >   	tmp |= (1 << 24) | (1 << 21) | (1 << 18);
> > -	intel_sbi_write(dev_priv, 0x216C, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x216C, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2080, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2080, SBI_MPHY);
> >   	tmp &= ~(7 << 13);
> >   	tmp |= (5 << 13);
> > -	intel_sbi_write(dev_priv, 0x2080, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2080, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2180, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2180, SBI_MPHY);
> >   	tmp &= ~(7 << 13);
> >   	tmp |= (5 << 13);
> > -	intel_sbi_write(dev_priv, 0x2180, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2180, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x208C, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x208C, SBI_MPHY);
> >   	tmp &= ~0xFF;
> >   	tmp |= 0x1C;
> > -	intel_sbi_write(dev_priv, 0x208C, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x208C, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x218C, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x218C, SBI_MPHY);
> >   	tmp &= ~0xFF;
> >   	tmp |= 0x1C;
> > -	intel_sbi_write(dev_priv, 0x218C, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x218C, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2098, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2098, SBI_MPHY);
> >   	tmp &= ~(0xFF << 16);
> >   	tmp |= (0x1C << 16);
> > -	intel_sbi_write(dev_priv, 0x2098, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2098, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x2198, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x2198, SBI_MPHY);
> >   	tmp &= ~(0xFF << 16);
> >   	tmp |= (0x1C << 16);
> > -	intel_sbi_write(dev_priv, 0x2198, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x2198, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x20C4, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x20C4, SBI_MPHY);
> >   	tmp |= (1 << 27);
> > -	intel_sbi_write(dev_priv, 0x20C4, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x20C4, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x21C4, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x21C4, SBI_MPHY);
> >   	tmp |= (1 << 27);
> > -	intel_sbi_write(dev_priv, 0x21C4, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x21C4, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x20EC, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x20EC, SBI_MPHY);
> >   	tmp &= ~(0xF << 28);
> >   	tmp |= (4 << 28);
> > -	intel_sbi_write(dev_priv, 0x20EC, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x20EC, tmp, SBI_MPHY);
> > -	tmp = intel_sbi_read(dev_priv, 0x21EC, SBI_MPHY);
> > +	tmp = intel_sbi_read(i915, 0x21EC, SBI_MPHY);
> >   	tmp &= ~(0xF << 28);
> >   	tmp |= (4 << 28);
> > -	intel_sbi_write(dev_priv, 0x21EC, tmp, SBI_MPHY);
> > +	intel_sbi_write(i915, 0x21EC, tmp, SBI_MPHY);
> >   }
> >   /* Implements 3 different sequences from BSpec chapter "Display iCLK
> > @@ -9000,69 +9000,69 @@ static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv)
> >    * - Sequence to enable CLKOUT_DP without spread
> >    * - Sequence to enable CLKOUT_DP for FDI usage and configure PCH FDI I/O
> >    */
> > -static void lpt_enable_clkout_dp(struct drm_i915_private *dev_priv,
> > +static void lpt_enable_clkout_dp(struct drm_i915_private *i915,
> >   				 bool with_spread, bool with_fdi)
> >   {
> >   	u32 reg, tmp;
> >   	if (WARN(with_fdi && !with_spread, "FDI requires downspread\n"))
> >   		with_spread = true;
> > -	if (WARN(HAS_PCH_LPT_LP(dev_priv) &&
> > +	if (WARN(HAS_PCH_LPT_LP(i915) &&
> >   	    with_fdi, "LP PCH doesn't have FDI\n"))
> >   		with_fdi = false;
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> > -	tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> > +	tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> >   	tmp &= ~SBI_SSCCTL_DISABLE;
> >   	tmp |= SBI_SSCCTL_PATHALT;
> > -	intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> >   	udelay(24);
> >   	if (with_spread) {
> > -		tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> > +		tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> >   		tmp &= ~SBI_SSCCTL_PATHALT;
> > -		intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> > +		intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> >   		if (with_fdi) {
> > -			lpt_reset_fdi_mphy(dev_priv);
> > -			lpt_program_fdi_mphy(dev_priv);
> > +			lpt_reset_fdi_mphy(i915);
> > +			lpt_program_fdi_mphy(i915);
> >   		}
> >   	}
> > -	reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0;
> > -	tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK);
> > +	reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0;
> > +	tmp = intel_sbi_read(i915, reg, SBI_ICLK);
> >   	tmp |= SBI_GEN0_CFG_BUFFENABLE_DISABLE;
> > -	intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK);
> > +	intel_sbi_write(i915, reg, tmp, SBI_ICLK);
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   }
> >   /* Sequence to disable CLKOUT_DP */
> > -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv)
> > +void lpt_disable_clkout_dp(struct drm_i915_private *i915)
> >   {
> >   	u32 reg, tmp;
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> > -	reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0;
> > -	tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK);
> > +	reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0;
> > +	tmp = intel_sbi_read(i915, reg, SBI_ICLK);
> >   	tmp &= ~SBI_GEN0_CFG_BUFFENABLE_DISABLE;
> > -	intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK);
> > +	intel_sbi_write(i915, reg, tmp, SBI_ICLK);
> > -	tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK);
> > +	tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK);
> >   	if (!(tmp & SBI_SSCCTL_DISABLE)) {
> >   		if (!(tmp & SBI_SSCCTL_PATHALT)) {
> >   			tmp |= SBI_SSCCTL_PATHALT;
> > -			intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> > +			intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> >   			udelay(32);
> >   		}
> >   		tmp |= SBI_SSCCTL_DISABLE;
> > -		intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK);
> > +		intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK);
> >   	}
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   }
> >   #define BEND_IDX(steps) ((50 + (steps)) / 5)
> > @@ -9097,7 +9097,7 @@ static const u16 sscdivintphase[] = {
> >    * < 0 slow down the clock, > 0 speed up the clock, 0 == no bend (135MHz)
> >    * change in clock period = -(steps / 10) * 5.787 ps
> >    */
> > -static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps)
> > +static void lpt_bend_clkout_dp(struct drm_i915_private *i915, int steps)
> >   {
> >   	u32 tmp;
> >   	int idx = BEND_IDX(steps);
> > @@ -9108,25 +9108,25 @@ static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps)
> >   	if (WARN_ON(idx >= ARRAY_SIZE(sscdivintphase)))
> >   		return;
> > -	mutex_lock(&dev_priv->sb_lock);
> > +	mutex_lock(&i915->sb_lock);
> >   	if (steps % 10 != 0)
> >   		tmp = 0xAAAAAAAB;
> >   	else
> >   		tmp = 0x00000000;
> > -	intel_sbi_write(dev_priv, SBI_SSCDITHPHASE, tmp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCDITHPHASE, tmp, SBI_ICLK);
> > -	tmp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE, SBI_ICLK);
> > +	tmp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE, SBI_ICLK);
> >   	tmp &= 0xffff0000;
> >   	tmp |= sscdivintphase[idx];
> > -	intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK);
> > +	intel_sbi_write(i915, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK);
> > -	mutex_unlock(&dev_priv->sb_lock);
> > +	mutex_unlock(&i915->sb_lock);
> >   }
> >   #undef BEND_IDX
> > -static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv)
> > +static bool spll_uses_pch_ssc(struct drm_i915_private *i915)
> >   {
> >   	u32 fuse_strap = I915_READ(FUSE_STRAP);
> >   	u32 ctl = I915_READ(SPLL_CTL);
> > @@ -9138,14 +9138,14 @@ static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv)
> >   	    (fuse_strap & HSW_CPU_SSC_ENABLE) == 0)
> >   		return true;
> > -	if (IS_BROADWELL(dev_priv) &&
> > +	if (IS_BROADWELL(i915) &&
> >   	    (ctl & SPLL_REF_MASK) == SPLL_REF_PCH_SSC_BDW)
> >   		return true;
> >   	return false;
> >   }
> > -static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> > +static bool wrpll_uses_pch_ssc(struct drm_i915_private *i915,
> >   			       enum intel_dpll_id id)
> >   {
> >   	u32 fuse_strap = I915_READ(FUSE_STRAP);
> > @@ -9157,7 +9157,7 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> >   	if ((ctl & WRPLL_REF_MASK) == WRPLL_REF_PCH_SSC)
> >   		return true;
> > -	if ((IS_BROADWELL(dev_priv) || IS_HSW_ULT(dev_priv)) &&
> > +	if ((IS_BROADWELL(i915) || IS_HSW_ULT(i915)) &&
> >   	    (ctl & WRPLL_REF_MASK) == WRPLL_REF_MUXED_SSC_BDW &&
> >   	    (fuse_strap & HSW_CPU_SSC_ENABLE) == 0)
> >   		return true;
> > @@ -9165,13 +9165,13 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv,
> >   	return false;
> >   }
> > -static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> > +static void lpt_init_pch_refclk(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> >   	bool pch_ssc_in_use = false;
> >   	bool has_fdi = false;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		switch (encoder->type) {
> >   		case INTEL_OUTPUT_ANALOG:
> >   			has_fdi = true;
> > @@ -9196,17 +9196,17 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   	 * clock hierarchy. That would also allow us to do
> >   	 * clock bending finally.
> >   	 */
> > -	if (spll_uses_pch_ssc(dev_priv)) {
> > +	if (spll_uses_pch_ssc(i915)) {
> >   		DRM_DEBUG_KMS("SPLL using PCH SSC\n");
> >   		pch_ssc_in_use = true;
> >   	}
> > -	if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL1)) {
> > +	if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL1)) {
> >   		DRM_DEBUG_KMS("WRPLL1 using PCH SSC\n");
> >   		pch_ssc_in_use = true;
> >   	}
> > -	if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL2)) {
> > +	if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL2)) {
> >   		DRM_DEBUG_KMS("WRPLL2 using PCH SSC\n");
> >   		pch_ssc_in_use = true;
> >   	}
> > @@ -9215,28 +9215,28 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv)
> >   		return;
> >   	if (has_fdi) {
> > -		lpt_bend_clkout_dp(dev_priv, 0);
> > -		lpt_enable_clkout_dp(dev_priv, true, true);
> > +		lpt_bend_clkout_dp(i915, 0);
> > +		lpt_enable_clkout_dp(i915, true, true);
> >   	} else {
> > -		lpt_disable_clkout_dp(dev_priv);
> > +		lpt_disable_clkout_dp(i915);
> >   	}
> >   }
> >   /*
> >    * Initialize reference clocks when the driver loads
> >    */
> > -void intel_init_pch_refclk(struct drm_i915_private *dev_priv)
> > +void intel_init_pch_refclk(struct drm_i915_private *i915)
> >   {
> > -	if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))
> > -		ironlake_init_pch_refclk(dev_priv);
> > -	else if (HAS_PCH_LPT(dev_priv))
> > -		lpt_init_pch_refclk(dev_priv);
> > +	if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))
> > +		ironlake_init_pch_refclk(i915);
> > +	else if (HAS_PCH_LPT(i915))
> > +		lpt_init_pch_refclk(i915);
> >   }
> >   static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > @@ -9280,11 +9280,11 @@ static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 val = 0;
> > -	if (IS_HASWELL(dev_priv) && crtc_state->dither)
> > +	if (IS_HASWELL(i915) && crtc_state->dither)
> >   		val |= (PIPECONF_DITHER_EN | PIPECONF_DITHER_TYPE_SP);
> >   	if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
> > @@ -9299,7 +9299,7 @@ static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state)
> >   static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 val = 0;
> >   	switch (crtc_state->pipe_bpp) {
> > @@ -9331,7 +9331,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
> >   		val |= PIPEMISC_YUV420_ENABLE |
> >   			PIPEMISC_YUV420_MODE_FULL_BLEND;
> > -	if (INTEL_GEN(dev_priv) >= 11 &&
> > +	if (INTEL_GEN(i915) >= 11 &&
> >   	    (crtc_state->active_planes & ~(icl_hdr_plane_mask() |
> >   					   BIT(PLANE_CURSOR))) == 0)
> >   		val |= PIPEMISC_HDR_MODE_PRECISION;
> > @@ -9341,7 +9341,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state)
> >   int bdw_get_pipemisc_bpp(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 tmp;
> >   	tmp = I915_READ(PIPEMISC(crtc->pipe));
> > @@ -9381,17 +9381,17 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> >   				  struct intel_crtc_state *crtc_state,
> >   				  struct dpll *reduced_clock)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 dpll, fp, fp2;
> >   	int factor;
> >   	/* Enable autotuning of the PLL clock (if permissible) */
> >   	factor = 21;
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if ((intel_panel_use_ssc(dev_priv) &&
> > -		     dev_priv->vbt.lvds_ssc_freq == 100000) ||
> > -		    (HAS_PCH_IBX(dev_priv) &&
> > -		     intel_is_dual_link_lvds(dev_priv)))
> > +		if ((intel_panel_use_ssc(i915) &&
> > +		     i915->vbt.lvds_ssc_freq == 100000) ||
> > +		    (HAS_PCH_IBX(i915) &&
> > +		     intel_is_dual_link_lvds(i915)))
> >   			factor = 25;
> >   	} else if (crtc_state->sdvo_tv_clock) {
> >   		factor = 20;
> > @@ -9442,7 +9442,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> >   	 * clear if it''s a win or loss power wise. No point in doing
> >   	 * this on ILK at all since it has a fixed DPLL<->pipe mapping.
> >   	 */
> > -	if (INTEL_INFO(dev_priv)->num_pipes == 3 &&
> > +	if (INTEL_INFO(i915)->num_pipes == 3 &&
> >   	    intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG))
> >   		dpll |= DPLL_SDVO_HIGH_SPEED;
> > @@ -9467,7 +9467,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> >   	}
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) &&
> > -	    intel_panel_use_ssc(dev_priv))
> > +	    intel_panel_use_ssc(i915))
> >   		dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN;
> >   	else
> >   		dpll |= PLL_REF_INPUT_DREFCLK;
> > @@ -9482,7 +9482,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc,
> >   static int ironlake_crtc_compute_clock(struct intel_crtc *crtc,
> >   				       struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct intel_limit *limit;
> >   	int refclk = 120000;
> > @@ -9494,13 +9494,13 @@ static int ironlake_crtc_compute_clock(struct intel_crtc *crtc,
> >   		return 0;
> >   	if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) {
> > -		if (intel_panel_use_ssc(dev_priv)) {
> > +		if (intel_panel_use_ssc(i915)) {
> >   			DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n",
> > -				      dev_priv->vbt.lvds_ssc_freq);
> > -			refclk = dev_priv->vbt.lvds_ssc_freq;
> > +				      i915->vbt.lvds_ssc_freq);
> > +			refclk = i915->vbt.lvds_ssc_freq;
> >   		}
> > -		if (intel_is_dual_link_lvds(dev_priv)) {
> > +		if (intel_is_dual_link_lvds(i915)) {
> >   			if (refclk == 100000)
> >   				limit = &intel_limits_ironlake_dual_lvds_100m;
> >   			else
> > @@ -9537,7 +9537,7 @@ static void intel_pch_transcoder_get_m_n(struct intel_crtc *crtc,
> >   					 struct intel_link_m_n *m_n)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	m_n->link_m = I915_READ(PCH_TRANS_LINK_M1(pipe));
> > @@ -9554,10 +9554,10 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc,
> >   					 struct intel_link_m_n *m_n,
> >   					 struct intel_link_m_n *m2_n2)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> > -	if (INTEL_GEN(dev_priv) >= 5) {
> > +	if (INTEL_GEN(i915) >= 5) {
> >   		m_n->link_m = I915_READ(PIPE_LINK_M1(transcoder));
> >   		m_n->link_n = I915_READ(PIPE_LINK_N1(transcoder));
> >   		m_n->gmch_m = I915_READ(PIPE_DATA_M1(transcoder))
> > @@ -9566,7 +9566,7 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc,
> >   		m_n->tu = ((I915_READ(PIPE_DATA_M1(transcoder))
> >   			    & TU_SIZE_MASK) >> TU_SIZE_SHIFT) + 1;
> > -		if (m2_n2 && transcoder_has_m2_n2(dev_priv, transcoder)) {
> > +		if (m2_n2 && transcoder_has_m2_n2(i915, transcoder)) {
> >   			m2_n2->link_m = I915_READ(PIPE_LINK_M2(transcoder));
> >   			m2_n2->link_n =	I915_READ(PIPE_LINK_N2(transcoder));
> >   			m2_n2->gmch_m =	I915_READ(PIPE_DATA_M2(transcoder))
> > @@ -9608,7 +9608,7 @@ static void skylake_get_pfit_config(struct intel_crtc *crtc,
> >   				    struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc_scaler_state *scaler_state = &pipe_config->scaler_state;
> >   	u32 ps_ctrl = 0;
> >   	int id = -1;
> > @@ -9640,7 +9640,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
> >   				 struct intel_initial_plane_config *plane_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_plane *plane = to_intel_plane(crtc->base.primary);
> >   	enum plane_id plane_id = plane->id;
> >   	enum pipe pipe;
> > @@ -9667,12 +9667,12 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
> >   	val = I915_READ(PLANE_CTL(pipe, plane_id));
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		pixel_format = val & ICL_PLANE_CTL_FORMAT_MASK;
> >   	else
> >   		pixel_format = val & PLANE_CTL_FORMAT_MASK;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> >   		alpha = I915_READ(PLANE_COLOR_CTL(pipe, plane_id));
> >   		alpha &= PLANE_COLOR_ALPHA_MASK;
> >   	} else {
> > @@ -9729,7 +9729,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc,
> >   		break;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 10 &&
> > +	if (INTEL_GEN(i915) >= 10 &&
> >   	    val & PLANE_CTL_FLIP_HORIZONTAL)
> >   		plane_config->rotation |= DRM_MODE_REFLECT_X;
> > @@ -9766,7 +9766,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc,
> >   				     struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 tmp;
> >   	tmp = I915_READ(PF_CTL(crtc->pipe));
> > @@ -9779,7 +9779,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc,
> >   		/* We currently do not free assignements of panel fitters on
> >   		 * ivb/hsw (since we don't use the higher upscaling modes which
> >   		 * differentiates them) so just WARN about this case for now. */
> > -		if (IS_GEN(dev_priv, 7)) {
> > +		if (IS_GEN(i915, 7)) {
> >   			WARN_ON((tmp & PF_PIPE_SEL_MASK_IVB) !=
> >   				PF_PIPE_SEL_IVB(crtc->pipe));
> >   		}
> > @@ -9790,14 +9790,14 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> >   				     struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	u32 tmp;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -9850,7 +9850,7 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> >   		ironlake_get_fdi_m_n_config(crtc, pipe_config);
> > -		if (HAS_PCH_IBX(dev_priv)) {
> > +		if (HAS_PCH_IBX(i915)) {
> >   			/*
> >   			 * The pipe->pch transcoder and pch transcoder->pll
> >   			 * mapping is fixed.
> > @@ -9865,10 +9865,10 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> >   		}
> >   		pipe_config->shared_dpll =
> > -			intel_get_shared_dpll_by_id(dev_priv, pll_id);
> > +			intel_get_shared_dpll_by_id(i915, pll_id);
> >   		pll = pipe_config->shared_dpll;
> > -		WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll,
> > +		WARN_ON(!pll->info->funcs->get_hw_state(i915, pll,
> >   						&pipe_config->dpll_hw_state));
> >   		tmp = pipe_config->dpll_hw_state.dpll;
> > @@ -9889,19 +9889,19 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> >   static int haswell_crtc_compute_clock(struct intel_crtc *crtc,
> >   				      struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_atomic_state *state =
> >   		to_intel_atomic_state(crtc_state->base.state);
> >   	if (!intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) ||
> > -	    INTEL_GEN(dev_priv) >= 11) {
> > +	    INTEL_GEN(i915) >= 11) {
> >   		struct intel_encoder *encoder =
> >   			intel_get_crtc_new_encoder(state, crtc_state);
> > @@ -9915,7 +9915,7 @@ static int haswell_crtc_compute_clock(struct intel_crtc *crtc,
> >   	return 0;
> >   }
> > -static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv,
> > +static void cannonlake_get_ddi_pll(struct drm_i915_private *i915,
> >   				   enum port port,
> >   				   struct intel_crtc_state *pipe_config)
> >   {
> > @@ -9928,10 +9928,10 @@ static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv,
> >   	if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL2))
> >   		return;
> > -	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> >   }
> > -static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv,
> > +static void icelake_get_ddi_pll(struct drm_i915_private *i915,
> >   				enum port port,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > @@ -9939,21 +9939,21 @@ static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv,
> >   	u32 temp;
> >   	/* TODO: TBT pll not implemented. */
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > +	if (intel_port_is_combophy(i915, port)) {
> >   		temp = I915_READ(DPCLKA_CFGCR0_ICL) &
> >   		       DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port);
> >   		id = temp >> DPCLKA_CFGCR0_DDI_CLK_SEL_SHIFT(port);
> > -	} else if (intel_port_is_tc(dev_priv, port)) {
> > -		id = icl_tc_port_to_pll_id(intel_port_to_tc(dev_priv, port));
> > +	} else if (intel_port_is_tc(i915, port)) {
> > +		id = icl_tc_port_to_pll_id(intel_port_to_tc(i915, port));
> >   	} else {
> >   		WARN(1, "Invalid port %x\n", port);
> >   		return;
> >   	}
> > -	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> >   }
> > -static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv,
> > +static void bxt_get_ddi_pll(struct drm_i915_private *i915,
> >   				enum port port,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > @@ -9974,10 +9974,10 @@ static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> >   }
> > -static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv,
> > +static void skylake_get_ddi_pll(struct drm_i915_private *i915,
> >   				enum port port,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > @@ -9990,10 +9990,10 @@ static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv,
> >   	if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL3))
> >   		return;
> > -	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> >   }
> > -static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv,
> > +static void haswell_get_ddi_pll(struct drm_i915_private *i915,
> >   				enum port port,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > @@ -10026,7 +10026,7 @@ static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id);
> >   }
> >   static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> > @@ -10035,7 +10035,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> >   				     intel_wakeref_t *wakerefs)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum intel_display_power_domain power_domain;
> >   	unsigned long panel_transcoder_mask = 0;
> >   	unsigned long enabled_panel_transcoders = 0;
> > @@ -10043,11 +10043,11 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> >   	intel_wakeref_t wf;
> >   	u32 tmp;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		panel_transcoder_mask |=
> >   			BIT(TRANSCODER_DSI_0) | BIT(TRANSCODER_DSI_1);
> > -	if (HAS_TRANSCODER_EDP(dev_priv))
> > +	if (HAS_TRANSCODER_EDP(i915))
> >   		panel_transcoder_mask |= BIT(TRANSCODER_EDP);
> >   	/*
> > @@ -10062,7 +10062,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> >   	 */
> >   	for_each_set_bit(panel_transcoder,
> >   			 &panel_transcoder_mask,
> > -			 ARRAY_SIZE(INTEL_INFO(dev_priv)->trans_offsets)) {
> > +			 ARRAY_SIZE(INTEL_INFO(i915)->trans_offsets)) {
> >   		bool force_thru = false;
> >   		enum pipe trans_pipe;
> > @@ -10113,7 +10113,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc,
> >   	power_domain = POWER_DOMAIN_TRANSCODER(pipe_config->cpu_transcoder);
> >   	WARN_ON(*power_domain_mask & BIT_ULL(power_domain));
> > -	wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wf = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wf)
> >   		return false;
> > @@ -10131,7 +10131,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> >   					 intel_wakeref_t *wakerefs)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum intel_display_power_domain power_domain;
> >   	enum transcoder cpu_transcoder;
> >   	intel_wakeref_t wf;
> > @@ -10147,7 +10147,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> >   		power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder);
> >   		WARN_ON(*power_domain_mask & BIT_ULL(power_domain));
> > -		wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +		wf = intel_display_power_get_if_enabled(i915, power_domain);
> >   		if (!wf)
> >   			continue;
> > @@ -10161,7 +10161,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> >   		 * registers/MIPI[BXT]. We can break out here early, since we
> >   		 * need the same DSI PLL to be enabled for both DSI ports.
> >   		 */
> > -		if (!bxt_dsi_pll_is_enabled(dev_priv))
> > +		if (!bxt_dsi_pll_is_enabled(i915))
> >   			break;
> >   		/* XXX: this works for video mode only */
> > @@ -10183,7 +10183,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc,
> >   static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> >   				       struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll;
> >   	enum port port;
> >   	u32 tmp;
> > @@ -10192,20 +10192,20 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> >   	port = (tmp & TRANS_DDI_PORT_MASK) >> TRANS_DDI_PORT_SHIFT;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		icelake_get_ddi_pll(dev_priv, port, pipe_config);
> > -	else if (IS_CANNONLAKE(dev_priv))
> > -		cannonlake_get_ddi_pll(dev_priv, port, pipe_config);
> > -	else if (IS_GEN9_BC(dev_priv))
> > -		skylake_get_ddi_pll(dev_priv, port, pipe_config);
> > -	else if (IS_GEN9_LP(dev_priv))
> > -		bxt_get_ddi_pll(dev_priv, port, pipe_config);
> > +	if (INTEL_GEN(i915) >= 11)
> > +		icelake_get_ddi_pll(i915, port, pipe_config);
> > +	else if (IS_CANNONLAKE(i915))
> > +		cannonlake_get_ddi_pll(i915, port, pipe_config);
> > +	else if (IS_GEN9_BC(i915))
> > +		skylake_get_ddi_pll(i915, port, pipe_config);
> > +	else if (IS_GEN9_LP(i915))
> > +		bxt_get_ddi_pll(i915, port, pipe_config);
> >   	else
> > -		haswell_get_ddi_pll(dev_priv, port, pipe_config);
> > +		haswell_get_ddi_pll(i915, port, pipe_config);
> >   	pll = pipe_config->shared_dpll;
> >   	if (pll) {
> > -		WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll,
> > +		WARN_ON(!pll->info->funcs->get_hw_state(i915, pll,
> >   						&pipe_config->dpll_hw_state));
> >   	}
> > @@ -10214,7 +10214,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> >   	 * DDI E. So just check whether this pipe is wired to DDI E and whether
> >   	 * the PCH transcoder is on.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 9 &&
> > +	if (INTEL_GEN(i915) < 9 &&
> >   	    (port == PORT_E) && I915_READ(LPT_TRANSCONF) & TRANS_ENABLE) {
> >   		pipe_config->has_pch_encoder = true;
> > @@ -10229,7 +10229,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc,
> >   static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   				    struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	intel_wakeref_t wakerefs[POWER_DOMAIN_NUM], wf;
> >   	enum intel_display_power_domain power_domain;
> >   	u64 power_domain_mask;
> > @@ -10238,7 +10238,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   	intel_crtc_init_scalers(crtc, pipe_config);
> >   	power_domain = POWER_DOMAIN_PIPE(crtc->pipe);
> > -	wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wf = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wf)
> >   		return false;
> > @@ -10250,7 +10250,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   	active = hsw_get_transcoder_state(crtc, pipe_config,
> >   					  &power_domain_mask, wakerefs);
> > -	if (IS_GEN9_LP(dev_priv) &&
> > +	if (IS_GEN9_LP(i915) &&
> >   	    bxt_get_dsi_transcoder_state(crtc, pipe_config,
> >   					 &power_domain_mask, wakerefs)) {
> >   		WARN_ON(active);
> > @@ -10261,7 +10261,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   		goto out;
> >   	if (!transcoder_is_dsi(pipe_config->cpu_transcoder) ||
> > -	    INTEL_GEN(dev_priv) >= 11) {
> > +	    INTEL_GEN(i915) >= 11) {
> >   		haswell_get_ddi_port_state(crtc, pipe_config);
> >   		intel_get_pipe_timings(crtc, pipe_config);
> >   	}
> > @@ -10273,7 +10273,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   	pipe_config->csc_mode = I915_READ(PIPE_CSC_MODE(crtc->pipe));
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		u32 tmp = I915_READ(SKL_BOTTOM_COLOR(crtc->pipe));
> >   		if (tmp & SKL_BOTTOM_COLOR_GAMMA_ENABLE)
> > @@ -10290,19 +10290,19 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   	power_domain = POWER_DOMAIN_PIPE_PANEL_FITTER(crtc->pipe);
> >   	WARN_ON(power_domain_mask & BIT_ULL(power_domain));
> > -	wf = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wf = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (wf) {
> >   		wakerefs[power_domain] = wf;
> >   		power_domain_mask |= BIT_ULL(power_domain);
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			skylake_get_pfit_config(crtc, pipe_config);
> >   		else
> >   			ironlake_get_pfit_config(crtc, pipe_config);
> >   	}
> >   	if (hsw_crtc_supports_ips(crtc)) {
> > -		if (IS_HASWELL(dev_priv))
> > +		if (IS_HASWELL(i915))
> >   			pipe_config->ips_enabled = I915_READ(IPS_CTL) & IPS_ENABLE;
> >   		else {
> >   			/*
> > @@ -10324,7 +10324,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   out:
> >   	for_each_power_domain(power_domain, power_domain_mask)
> > -		intel_display_power_put(dev_priv,
> > +		intel_display_power_put(i915,
> >   					power_domain, wakerefs[power_domain]);
> >   	return active;
> > @@ -10332,13 +10332,13 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc,
> >   static u32 intel_cursor_base(const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	const struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> >   	u32 base;
> > -	if (INTEL_INFO(dev_priv)->display.cursor_needs_physical)
> > +	if (INTEL_INFO(i915)->display.cursor_needs_physical)
> >   		base = obj->phys_handle->busaddr;
> >   	else
> >   		base = intel_plane_ggtt_offset(plane_state);
> > @@ -10346,7 +10346,7 @@ static u32 intel_cursor_base(const struct intel_plane_state *plane_state)
> >   	base += plane_state->color_plane[0].offset;
> >   	/* ILK+ do this automagically */
> > -	if (HAS_GMCH(dev_priv) &&
> > +	if (HAS_GMCH(i915) &&
> >   	    plane_state->base.rotation & DRM_MODE_ROTATE_180)
> >   		base += (plane_state->base.crtc_h *
> >   			 plane_state->base.crtc_w - 1) * fb->format->cpp[0];
> > @@ -10532,7 +10532,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	u32 cntl = 0, base = 0, pos = 0, size = 0;
> >   	unsigned long irqflags;
> > @@ -10549,7 +10549,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> >   		pos = intel_cursor_position(plane_state);
> >   	}
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	/* On these chipsets we can only modify the base/size/stride
> >   	 * whilst the cursor is disabled.
> > @@ -10570,7 +10570,7 @@ static void i845_update_cursor(struct intel_plane *plane,
> >   		I915_WRITE_FW(CURPOS(PIPE_A), pos);
> >   	}
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void i845_disable_cursor(struct intel_plane *plane,
> > @@ -10582,13 +10582,13 @@ static void i845_disable_cursor(struct intel_plane *plane,
> >   static bool i845_cursor_get_hw_state(struct intel_plane *plane,
> >   				     enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(PIPE_A);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -10596,7 +10596,7 @@ static bool i845_cursor_get_hw_state(struct intel_plane *plane,
> >   	*pipe = PIPE_A;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -10612,10 +10612,10 @@ i9xx_cursor_max_stride(struct intel_plane *plane,
> >   static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 cntl = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return cntl;
> >   	if (crtc_state->gamma_enable)
> > @@ -10624,7 +10624,7 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   	if (crtc_state->csc_enable)
> >   		cntl |= MCURSOR_PIPE_CSC_ENABLE;
> > -	if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> >   		cntl |= MCURSOR_PIPE_SELECT(crtc->pipe);
> >   	return cntl;
> > @@ -10633,11 +10633,11 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state,
> >   			   const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	u32 cntl = 0;
> > -	if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv))
> > +	if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915))
> >   		cntl |= MCURSOR_TRICKLE_FEED_DISABLE;
> >   	switch (plane_state->base.crtc_w) {
> > @@ -10663,7 +10663,7 @@ static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state,
> >   static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	int width = plane_state->base.crtc_w;
> >   	int height = plane_state->base.crtc_h;
> > @@ -10687,7 +10687,7 @@ static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state)
> >   	 * cursor is not rotated. Everything else requires square
> >   	 * cursors.
> >   	 */
> > -	if (HAS_CUR_FBC(dev_priv) &&
> > +	if (HAS_CUR_FBC(i915) &&
> >   	    plane_state->base.rotation & DRM_MODE_ROTATE_0) {
> >   		if (height < 8 || height > width)
> >   			return false;
> > @@ -10703,7 +10703,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state,
> >   			     struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	enum pipe pipe = plane->pipe;
> >   	int ret;
> > @@ -10743,7 +10743,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state,
> >   	 * display power well must be turned off and on again.
> >   	 * Refuse the put the cursor into that compromised position.
> >   	 */
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_C &&
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_C &&
> >   	    plane_state->base.visible && plane_state->base.crtc_x < 0) {
> >   		DRM_DEBUG_KMS("CHV cursor C not allowed to straddle the left screen edge\n");
> >   		return -EINVAL;
> > @@ -10758,7 +10758,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	u32 cntl = 0, base = 0, pos = 0, fbc_ctl = 0;
> >   	unsigned long irqflags;
> > @@ -10774,7 +10774,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> >   		pos = intel_cursor_position(plane_state);
> >   	}
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	/*
> >   	 * On some platforms writing CURCNTR first will also
> > @@ -10796,13 +10796,13 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> >   	 * the CURCNTR write arms the update.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		skl_write_cursor_wm(plane, crtc_state);
> >   	if (plane->cursor.base != base ||
> >   	    plane->cursor.size != fbc_ctl ||
> >   	    plane->cursor.cntl != cntl) {
> > -		if (HAS_CUR_FBC(dev_priv))
> > +		if (HAS_CUR_FBC(i915))
> >   			I915_WRITE_FW(CUR_FBC_CTL(pipe), fbc_ctl);
> >   		I915_WRITE_FW(CURCNTR(pipe), cntl);
> >   		I915_WRITE_FW(CURPOS(pipe), pos);
> > @@ -10816,7 +10816,7 @@ static void i9xx_update_cursor(struct intel_plane *plane,
> >   		I915_WRITE_FW(CURBASE(pipe), base);
> >   	}
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void i9xx_disable_cursor(struct intel_plane *plane,
> > @@ -10828,7 +10828,7 @@ static void i9xx_disable_cursor(struct intel_plane *plane,
> >   static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
> >   				     enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > @@ -10840,7 +10840,7 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
> >   	 * display power wells.
> >   	 */
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -10848,13 +10848,13 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane,
> >   	ret = val & MCURSOR_MODE;
> > -	if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> > +	if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> >   		*pipe = plane->pipe;
> >   	else
> >   		*pipe = (val & MCURSOR_PIPE_SELECT_MASK) >>
> >   			MCURSOR_PIPE_SELECT_SHIFT;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -10924,7 +10924,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector,
> >   	struct drm_encoder *encoder = &intel_encoder->base;
> >   	struct drm_crtc *crtc = NULL;
> >   	struct drm_device *dev = encoder->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_mode_config *config = &dev->mode_config;
> >   	struct drm_atomic_state *state = NULL, *restore_state = NULL;
> >   	struct drm_connector_state *connector_state;
> > @@ -11051,7 +11051,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector,
> >   	drm_atomic_state_put(state);
> >   	/* let the connector get through one full cycle before testing */
> > -	intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> > +	intel_wait_for_vblank(i915, intel_crtc->pipe);
> >   	return true;
> >   fail:
> > @@ -11096,14 +11096,14 @@ void intel_release_load_detect_pipe(struct drm_connector *connector,
> >   static int i9xx_pll_refclk(struct drm_device *dev,
> >   			   const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 dpll = pipe_config->dpll_hw_state.dpll;
> >   	if ((dpll & PLL_REF_INPUT_MASK) == PLLB_REF_INPUT_SPREADSPECTRUMIN)
> > -		return dev_priv->vbt.lvds_ssc_freq;
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > +		return i915->vbt.lvds_ssc_freq;
> > +	else if (HAS_PCH_SPLIT(i915))
> >   		return 120000;
> > -	else if (!IS_GEN(dev_priv, 2))
> > +	else if (!IS_GEN(i915, 2))
> >   		return 96000;
> >   	else
> >   		return 48000;
> > @@ -11114,7 +11114,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int pipe = pipe_config->cpu_transcoder;
> >   	u32 dpll = pipe_config->dpll_hw_state.dpll;
> >   	u32 fp;
> > @@ -11128,7 +11128,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> >   		fp = pipe_config->dpll_hw_state.fp1;
> >   	clock.m1 = (fp & FP_M1_DIV_MASK) >> FP_M1_DIV_SHIFT;
> > -	if (IS_PINEVIEW(dev_priv)) {
> > +	if (IS_PINEVIEW(i915)) {
> >   		clock.n = ffs((fp & FP_N_PINEVIEW_DIV_MASK) >> FP_N_DIV_SHIFT) - 1;
> >   		clock.m2 = (fp & FP_M2_PINEVIEW_DIV_MASK) >> FP_M2_DIV_SHIFT;
> >   	} else {
> > @@ -11136,8 +11136,8 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> >   		clock.m2 = (fp & FP_M2_DIV_MASK) >> FP_M2_DIV_SHIFT;
> >   	}
> > -	if (!IS_GEN(dev_priv, 2)) {
> > -		if (IS_PINEVIEW(dev_priv))
> > +	if (!IS_GEN(i915, 2)) {
> > +		if (IS_PINEVIEW(i915))
> >   			clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK_PINEVIEW) >>
> >   				DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW);
> >   		else
> > @@ -11159,12 +11159,12 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc,
> >   			return;
> >   		}
> > -		if (IS_PINEVIEW(dev_priv))
> > +		if (IS_PINEVIEW(i915))
> >   			port_clock = pnv_calc_dpll_params(refclk, &clock);
> >   		else
> >   			port_clock = i9xx_calc_dpll_params(refclk, &clock);
> >   	} else {
> > -		u32 lvds = IS_I830(dev_priv) ? 0 : I915_READ(LVDS);
> > +		u32 lvds = IS_I830(i915) ? 0 : I915_READ(LVDS);
> >   		bool is_lvds = (pipe == 1) && (lvds & LVDS_PORT_EN);
> >   		if (is_lvds) {
> > @@ -11221,7 +11221,7 @@ int intel_dotclock_calculate(int link_freq,
> >   static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> >   				   struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	/* read out port_clock from the DPLL */
> >   	i9xx_crtc_clock_get(crtc, pipe_config);
> > @@ -11232,7 +11232,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> >   	 * Calculate one based on the FDI configuration.
> >   	 */
> >   	pipe_config->base.adjusted_mode.crtc_clock =
> > -		intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config),
> > +		intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config),
> >   					 &pipe_config->fdi_m_n);
> >   }
> > @@ -11240,7 +11240,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc,
> >   struct drm_display_mode *
> >   intel_encoder_current_mode(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc_state *crtc_state;
> >   	struct drm_display_mode *mode;
> >   	struct intel_crtc *crtc;
> > @@ -11249,7 +11249,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder)
> >   	if (!encoder->get_hw_state(encoder, &pipe))
> >   		return NULL;
> > -	crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	mode = kzalloc(sizeof(*mode), GFP_KERNEL);
> >   	if (!mode)
> > @@ -11263,7 +11263,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder)
> >   	crtc_state->base.crtc = &crtc->base;
> > -	if (!dev_priv->display.get_pipe_config(crtc, crtc_state)) {
> > +	if (!i915->display.get_pipe_config(crtc, crtc_state)) {
> >   		kfree(crtc_state);
> >   		kfree(mode);
> >   		return NULL;
> > @@ -11337,7 +11337,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct intel_plane *plane = to_intel_plane(plane_state->plane);
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	bool mode_changed = needs_modeset(crtc_state);
> >   	bool was_crtc_enabled = old_crtc_state->base.active;
> >   	bool is_crtc_enabled = crtc_state->active;
> > @@ -11345,7 +11345,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> >   	struct drm_framebuffer *fb = plane_state->fb;
> >   	int ret;
> > -	if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_CURSOR) {
> > +	if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_CURSOR) {
> >   		ret = skl_update_scaler_plane(
> >   			to_intel_crtc_state(crtc_state),
> >   			to_intel_plane_state(plane_state));
> > @@ -11395,14 +11395,14 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> >   			 turn_off, turn_on, mode_changed);
> >   	if (turn_on) {
> > -		if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> > +		if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> >   			pipe_config->update_wm_pre = true;
> >   		/* must disable cxsr around plane enable/disable */
> >   		if (plane->id != PLANE_CURSOR)
> >   			pipe_config->disable_cxsr = true;
> >   	} else if (turn_off) {
> > -		if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv))
> > +		if (INTEL_GEN(i915) < 5 && !IS_G4X(i915))
> >   			pipe_config->update_wm_post = true;
> >   		/* must disable cxsr around plane enable/disable */
> > @@ -11410,7 +11410,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> >   			pipe_config->disable_cxsr = true;
> >   	} else if (intel_wm_need_update(to_intel_plane_state(plane->base.state),
> >   					to_intel_plane_state(plane_state))) {
> > -		if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) {
> > +		if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) {
> >   			/* FIXME bollocks */
> >   			pipe_config->update_wm_pre = true;
> >   			pipe_config->update_wm_post = true;
> > @@ -11454,8 +11454,8 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat
> >   	 * plane, not only sprite plane.
> >   	 */
> >   	if (plane->id != PLANE_CURSOR &&
> > -	    (IS_GEN_RANGE(dev_priv, 5, 6) ||
> > -	     IS_IVYBRIDGE(dev_priv)) &&
> > +	    (IS_GEN_RANGE(i915, 5, 6) ||
> > +	     IS_IVYBRIDGE(i915)) &&
> >   	    (turn_on || (!needs_scaling(old_plane_state) &&
> >   			 needs_scaling(to_intel_plane_state(plane_state)))))
> >   		pipe_config->disable_lp_wm = true;
> > @@ -11519,13 +11519,13 @@ static int icl_add_linked_planes(struct intel_atomic_state *state)
> >   static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_atomic_state *state = to_intel_atomic_state(crtc_state->base.state);
> >   	struct intel_plane *plane, *linked;
> >   	struct intel_plane_state *plane_state;
> >   	int i;
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return 0;
> >   	/*
> > @@ -11555,7 +11555,7 @@ static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state)
> >   		    !(crtc_state->nv12_planes & BIT(plane->id)))
> >   			continue;
> > -		for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, linked) {
> > +		for_each_intel_plane_on_crtc(&i915->drm, crtc, linked) {
> >   			if (!icl_is_nv12_y_plane(linked->id))
> >   				continue;
> > @@ -11602,21 +11602,21 @@ static bool c8_planes_changed(const struct intel_crtc_state *new_crtc_state)
> >   static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> >   				   struct drm_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct intel_crtc_state *pipe_config =
> >   		to_intel_crtc_state(crtc_state);
> >   	int ret;
> >   	bool mode_changed = needs_modeset(crtc_state);
> > -	if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv) &&
> > +	if (INTEL_GEN(i915) < 5 && !IS_G4X(i915) &&
> >   	    mode_changed && !crtc_state->active)
> >   		pipe_config->update_wm_post = true;
> >   	if (mode_changed && crtc_state->enable &&
> > -	    dev_priv->display.crtc_compute_clock &&
> > +	    i915->display.crtc_compute_clock &&
> >   	    !WARN_ON(pipe_config->shared_dpll)) {
> > -		ret = dev_priv->display.crtc_compute_clock(intel_crtc,
> > +		ret = i915->display.crtc_compute_clock(intel_crtc,
> >   							   pipe_config);
> >   		if (ret)
> >   			return ret;
> > @@ -11637,16 +11637,16 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> >   	}
> >   	ret = 0;
> > -	if (dev_priv->display.compute_pipe_wm) {
> > -		ret = dev_priv->display.compute_pipe_wm(pipe_config);
> > +	if (i915->display.compute_pipe_wm) {
> > +		ret = i915->display.compute_pipe_wm(pipe_config);
> >   		if (ret) {
> >   			DRM_DEBUG_KMS("Target pipe watermarks are invalid\n");
> >   			return ret;
> >   		}
> >   	}
> > -	if (dev_priv->display.compute_intermediate_wm) {
> > -		if (WARN_ON(!dev_priv->display.compute_pipe_wm))
> > +	if (i915->display.compute_intermediate_wm) {
> > +		if (WARN_ON(!i915->display.compute_pipe_wm))
> >   			return 0;
> >   		/*
> > @@ -11654,14 +11654,14 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> >   		 * old state and the new state.  We can program these
> >   		 * immediately.
> >   		 */
> > -		ret = dev_priv->display.compute_intermediate_wm(pipe_config);
> > +		ret = i915->display.compute_intermediate_wm(pipe_config);
> >   		if (ret) {
> >   			DRM_DEBUG_KMS("No valid intermediate pipe watermarks are possible\n");
> >   			return ret;
> >   		}
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		if (mode_changed || pipe_config->update_pipe)
> >   			ret = skl_update_scaler_crtc(pipe_config);
> > @@ -11671,11 +11671,11 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc,
> >   			ret = skl_check_pipe_max_pixel_rate(intel_crtc,
> >   							    pipe_config);
> >   		if (!ret)
> > -			ret = intel_atomic_setup_scalers(dev_priv, intel_crtc,
> > +			ret = intel_atomic_setup_scalers(i915, intel_crtc,
> >   							 pipe_config);
> >   	}
> > -	if (HAS_IPS(dev_priv))
> > +	if (HAS_IPS(i915))
> >   		pipe_config->ips_enabled = hsw_compute_ips_config(pipe_config);
> >   	return ret;
> > @@ -11752,16 +11752,16 @@ static int
> >   compute_baseline_pipe_bpp(struct intel_crtc *crtc,
> >   			  struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct drm_atomic_state *state = pipe_config->base.state;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_state *connector_state;
> >   	int bpp, i;
> > -	if ((IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) ||
> > -	    IS_CHERRYVIEW(dev_priv)))
> > +	if ((IS_G4X(i915) || IS_VALLEYVIEW(i915) ||
> > +	    IS_CHERRYVIEW(i915)))
> >   		bpp = 10*3;
> > -	else if (INTEL_GEN(dev_priv) >= 5)
> > +	else if (INTEL_GEN(i915) >= 5)
> >   		bpp = 12*3;
> >   	else
> >   		bpp = 8*3;
> > @@ -11807,13 +11807,13 @@ intel_dump_m_n_config(const struct intel_crtc_state *pipe_config,
> >   }
> >   static void
> > -intel_dump_infoframe(struct drm_i915_private *dev_priv,
> > +intel_dump_infoframe(struct drm_i915_private *i915,
> >   		     const union hdmi_infoframe *frame)
> >   {
> >   	if ((drm_debug & DRM_UT_KMS) == 0)
> >   		return;
> > -	hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, frame);
> > +	hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, frame);
> >   }
> >   #define OUTPUT_TYPE(x) [INTEL_OUTPUT_ ## x] = #x
> > @@ -11907,7 +11907,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> >   				   const char *context)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct intel_plane_state *plane_state;
> >   	struct intel_plane *plane;
> >   	char buf[64];
> > @@ -11953,13 +11953,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> >   		DRM_DEBUG_KMS("GCP: 0x%x\n", pipe_config->infoframes.gcp);
> >   	if (pipe_config->infoframes.enable &
> >   	    intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_AVI))
> > -		intel_dump_infoframe(dev_priv, &pipe_config->infoframes.avi);
> > +		intel_dump_infoframe(i915, &pipe_config->infoframes.avi);
> >   	if (pipe_config->infoframes.enable &
> >   	    intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_SPD))
> > -		intel_dump_infoframe(dev_priv, &pipe_config->infoframes.spd);
> > +		intel_dump_infoframe(i915, &pipe_config->infoframes.spd);
> >   	if (pipe_config->infoframes.enable &
> >   	    intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_VENDOR))
> > -		intel_dump_infoframe(dev_priv, &pipe_config->infoframes.hdmi);
> > +		intel_dump_infoframe(i915, &pipe_config->infoframes.hdmi);
> >   	DRM_DEBUG_KMS("requested mode:\n");
> >   	drm_mode_debug_printmodeline(&pipe_config->base.mode);
> > @@ -11971,13 +11971,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> >   		      pipe_config->pipe_src_w, pipe_config->pipe_src_h,
> >   		      pipe_config->pixel_rate);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		DRM_DEBUG_KMS("num_scalers: %d, scaler_users: 0x%x, scaler_id: %d\n",
> >   			      crtc->num_scalers,
> >   			      pipe_config->scaler_state.scaler_users,
> >   		              pipe_config->scaler_state.scaler_id);
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		DRM_DEBUG_KMS("gmch pfit: control: 0x%08x, ratios: 0x%08x, lvds border: 0x%08x\n",
> >   			      pipe_config->gmch_pfit.control,
> >   			      pipe_config->gmch_pfit.pgm_ratios,
> > @@ -11992,7 +11992,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config,
> >   	DRM_DEBUG_KMS("ips: %i, double wide: %i\n",
> >   		      pipe_config->ips_enabled, pipe_config->double_wide);
> > -	intel_dpll_dump_hw_state(dev_priv, &pipe_config->dpll_hw_state);
> > +	intel_dpll_dump_hw_state(i915, &pipe_config->dpll_hw_state);
> >   dump_planes:
> >   	if (!state)
> > @@ -12073,7 +12073,7 @@ static bool check_digital_port_conflicts(struct intel_atomic_state *state)
> >   static int
> >   clear_intel_crtc_state(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_crtc_state *saved_state;
> > @@ -12090,8 +12090,8 @@ clear_intel_crtc_state(struct intel_crtc_state *crtc_state)
> >   	saved_state->shared_dpll = crtc_state->shared_dpll;
> >   	saved_state->dpll_hw_state = crtc_state->dpll_hw_state;
> >   	saved_state->crc_enabled = crtc_state->crc_enabled;
> > -	if (IS_G4X(dev_priv) ||
> > -	    IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_G4X(i915) ||
> > +	    IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		saved_state->wm = crtc_state->wm;
> >   	/* Keep base drm_crtc_state intact, only clear our extended struct */
> > @@ -12316,7 +12316,7 @@ intel_compare_infoframe(const union hdmi_infoframe *a,
> >   }
> >   static void
> > -pipe_config_infoframe_err(struct drm_i915_private *dev_priv,
> > +pipe_config_infoframe_err(struct drm_i915_private *i915,
> >   			  bool adjust, const char *name,
> >   			  const union hdmi_infoframe *a,
> >   			  const union hdmi_infoframe *b)
> > @@ -12327,15 +12327,15 @@ pipe_config_infoframe_err(struct drm_i915_private *dev_priv,
> >   		drm_dbg(DRM_UT_KMS, "mismatch in %s infoframe", name);
> >   		drm_dbg(DRM_UT_KMS, "expected:");
> > -		hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, a);
> > +		hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, a);
> >   		drm_dbg(DRM_UT_KMS, "found");
> > -		hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, b);
> > +		hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, b);
> >   	} else {
> >   		drm_err("mismatch in %s infoframe", name);
> >   		drm_err("expected:");
> > -		hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, a);
> > +		hdmi_infoframe_log(KERN_ERR, i915->drm.dev, a);
> >   		drm_err("found");
> > -		hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, b);
> > +		hdmi_infoframe_log(KERN_ERR, i915->drm.dev, b);
> >   	}
> >   }
> > @@ -12357,17 +12357,17 @@ pipe_config_err(bool adjust, const char *name, const char *format, ...)
> >   	va_end(args);
> >   }
> > -static bool fastboot_enabled(struct drm_i915_private *dev_priv)
> > +static bool fastboot_enabled(struct drm_i915_private *i915)
> >   {
> >   	if (i915_modparams.fastboot != -1)
> >   		return i915_modparams.fastboot;
> >   	/* Enable fastboot by default on Skylake and newer */
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return true;
> >   	/* Enable fastboot by default on VLV and CHV */
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		return true;
> >   	/* Disabled by default on all others */
> > @@ -12375,7 +12375,7 @@ static bool fastboot_enabled(struct drm_i915_private *dev_priv)
> >   }
> >   static bool
> > -intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> > +intel_pipe_config_compare(struct drm_i915_private *i915,
> >   			  struct intel_crtc_state *current_config,
> >   			  struct intel_crtc_state *pipe_config,
> >   			  bool adjust)
> > @@ -12385,7 +12385,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   		(current_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED) &&
> >   		!(pipe_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED);
> > -	if (fixup_inherited && !fastboot_enabled(dev_priv)) {
> > +	if (fixup_inherited && !fastboot_enabled(i915)) {
> >   		DRM_DEBUG_KMS("initial modeset and fastboot not set\n");
> >   		ret = false;
> >   	}
> > @@ -12525,7 +12525,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   #define PIPE_CONF_CHECK_INFOFRAME(name) do { \
> >   	if (!intel_compare_infoframe(&current_config->infoframes.name, \
> >   				     &pipe_config->infoframes.name)) { \
> > -		pipe_config_infoframe_err(dev_priv, adjust, __stringify(name), \
> > +		pipe_config_infoframe_err(i915, adjust, __stringify(name), \
> >   					  &current_config->infoframes.name, \
> >   					  &pipe_config->infoframes.name); \
> >   		ret = false; \
> > @@ -12544,7 +12544,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   	PIPE_CONF_CHECK_I(lane_count);
> >   	PIPE_CONF_CHECK_X(lane_lat_optim_mask);
> > -	if (INTEL_GEN(dev_priv) < 8) {
> > +	if (INTEL_GEN(i915) < 8) {
> >   		PIPE_CONF_CHECK_M_N(dp_m_n);
> >   		if (current_config->has_drrs)
> > @@ -12571,8 +12571,8 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   	PIPE_CONF_CHECK_I(pixel_multiplier);
> >   	PIPE_CONF_CHECK_I(output_format);
> >   	PIPE_CONF_CHECK_BOOL(has_hdmi_sink);
> > -	if ((INTEL_GEN(dev_priv) < 8 && !IS_HASWELL(dev_priv)) ||
> > -	    IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if ((INTEL_GEN(i915) < 8 && !IS_HASWELL(i915)) ||
> > +	    IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		PIPE_CONF_CHECK_BOOL(limited_color_range);
> >   	PIPE_CONF_CHECK_BOOL(hdmi_scrambling);
> > @@ -12597,7 +12597,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   	PIPE_CONF_CHECK_X(gmch_pfit.control);
> >   	/* pfit ratios are autocomputed by the hw on gen4+ */
> > -	if (INTEL_GEN(dev_priv) < 4)
> > +	if (INTEL_GEN(i915) < 4)
> >   		PIPE_CONF_CHECK_X(gmch_pfit.pgm_ratios);
> >   	PIPE_CONF_CHECK_X(gmch_pfit.lvds_border_bits);
> > @@ -12621,7 +12621,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   		PIPE_CONF_CHECK_CLOCK_FUZZY(pixel_rate);
> >   		PIPE_CONF_CHECK_X(gamma_mode);
> > -		if (IS_CHERRYVIEW(dev_priv))
> > +		if (IS_CHERRYVIEW(i915))
> >   			PIPE_CONF_CHECK_X(cgm_mode);
> >   		else
> >   			PIPE_CONF_CHECK_X(csc_mode);
> > @@ -12667,7 +12667,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   	PIPE_CONF_CHECK_X(dsi_pll.ctrl);
> >   	PIPE_CONF_CHECK_X(dsi_pll.div);
> > -	if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5)
> > +	if (IS_G4X(i915) || INTEL_GEN(i915) >= 5)
> >   		PIPE_CONF_CHECK_I(pipe_bpp);
> >   	PIPE_CONF_CHECK_CLOCK_FUZZY(base.adjusted_mode.crtc_clock);
> > @@ -12694,11 +12694,11 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv,
> >   	return ret;
> >   }
> > -static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv,
> > +static void intel_pipe_config_sanity_check(struct drm_i915_private *i915,
> >   					   const struct intel_crtc_state *pipe_config)
> >   {
> >   	if (pipe_config->has_pch_encoder) {
> > -		int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config),
> > +		int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config),
> >   							    &pipe_config->fdi_m_n);
> >   		int dotclock = pipe_config->base.adjusted_mode.crtc_clock;
> > @@ -12715,7 +12715,7 @@ static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv,
> >   static void verify_wm_state(struct drm_crtc *crtc,
> >   			    struct drm_crtc_state *new_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct skl_hw_state {
> >   		struct skl_ddb_entry ddb_y[I915_MAX_PLANES];
> >   		struct skl_ddb_entry ddb_uv[I915_MAX_PLANES];
> > @@ -12727,9 +12727,9 @@ static void verify_wm_state(struct drm_crtc *crtc,
> >   	struct skl_ddb_entry *hw_ddb_entry, *sw_ddb_entry;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	const enum pipe pipe = intel_crtc->pipe;
> > -	int plane, level, max_level = ilk_wm_max_level(dev_priv);
> > +	int plane, level, max_level = ilk_wm_max_level(i915);
> > -	if (INTEL_GEN(dev_priv) < 9 || !new_state->active)
> > +	if (INTEL_GEN(i915) < 9 || !new_state->active)
> >   		return;
> >   	hw = kzalloc(sizeof(*hw), GFP_KERNEL);
> > @@ -12741,17 +12741,17 @@ static void verify_wm_state(struct drm_crtc *crtc,
> >   	skl_pipe_ddb_get_hw_state(intel_crtc, hw->ddb_y, hw->ddb_uv);
> > -	skl_ddb_get_hw_state(dev_priv, &hw->ddb);
> > -	sw_ddb = &dev_priv->wm.skl_hw.ddb;
> > +	skl_ddb_get_hw_state(i915, &hw->ddb);
> > +	sw_ddb = &i915->wm.skl_hw.ddb;
> > -	if (INTEL_GEN(dev_priv) >= 11 &&
> > +	if (INTEL_GEN(i915) >= 11 &&
> >   	    hw->ddb.enabled_slices != sw_ddb->enabled_slices)
> >   		DRM_ERROR("mismatch in DBUF Slices (expected %u, got %u)\n",
> >   			  sw_ddb->enabled_slices,
> >   			  hw->ddb.enabled_slices);
> >   	/* planes */
> > -	for_each_universal_plane(dev_priv, pipe, plane) {
> > +	for_each_universal_plane(i915, pipe, plane) {
> >   		struct skl_plane_wm *hw_plane_wm, *sw_plane_wm;
> >   		hw_plane_wm = &hw->wm.planes[plane];
> > @@ -12933,7 +12933,7 @@ verify_crtc_state(struct drm_crtc *crtc,
> >   		  struct drm_crtc_state *new_crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_encoder *encoder;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct intel_crtc_state *pipe_config, *sw_config;
> > @@ -12949,10 +12949,10 @@ verify_crtc_state(struct drm_crtc *crtc,
> >   	DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
> > -	active = dev_priv->display.get_pipe_config(intel_crtc, pipe_config);
> > +	active = i915->display.get_pipe_config(intel_crtc, pipe_config);
> >   	/* we keep both pipes enabled on 830 */
> > -	if (IS_I830(dev_priv))
> > +	if (IS_I830(i915))
> >   		active = new_crtc_state->active;
> >   	I915_STATE_WARN(new_crtc_state->active != active,
> > @@ -12984,10 +12984,10 @@ verify_crtc_state(struct drm_crtc *crtc,
> >   	if (!new_crtc_state->active)
> >   		return;
> > -	intel_pipe_config_sanity_check(dev_priv, pipe_config);
> > +	intel_pipe_config_sanity_check(i915, pipe_config);
> >   	sw_config = to_intel_crtc_state(new_crtc_state);
> > -	if (!intel_pipe_config_compare(dev_priv, sw_config,
> > +	if (!intel_pipe_config_compare(i915, sw_config,
> >   				       pipe_config, false)) {
> >   		I915_STATE_WARN(1, "pipe state doesn't match!\n");
> >   		intel_dump_pipe_config(pipe_config, NULL, "[hw state]");
> > @@ -13009,7 +13009,7 @@ intel_verify_planes(struct intel_atomic_state *state)
> >   }
> >   static void
> > -verify_single_dpll_state(struct drm_i915_private *dev_priv,
> > +verify_single_dpll_state(struct drm_i915_private *i915,
> >   			 struct intel_shared_dpll *pll,
> >   			 struct drm_crtc *crtc,
> >   			 struct drm_crtc_state *new_state)
> > @@ -13022,7 +13022,7 @@ verify_single_dpll_state(struct drm_i915_private *dev_priv,
> >   	DRM_DEBUG_KMS("%s\n", pll->info->name);
> > -	active = pll->info->funcs->get_hw_state(dev_priv, pll, &dpll_hw_state);
> > +	active = pll->info->funcs->get_hw_state(i915, pll, &dpll_hw_state);
> >   	if (!(pll->info->flags & INTEL_DPLL_ALWAYS_ON)) {
> >   		I915_STATE_WARN(!pll->on && pll->active_mask,
> > @@ -13068,12 +13068,12 @@ verify_shared_dpll_state(struct drm_device *dev, struct drm_crtc *crtc,
> >   			 struct drm_crtc_state *old_crtc_state,
> >   			 struct drm_crtc_state *new_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc_state *old_state = to_intel_crtc_state(old_crtc_state);
> >   	struct intel_crtc_state *new_state = to_intel_crtc_state(new_crtc_state);
> >   	if (new_state->shared_dpll)
> > -		verify_single_dpll_state(dev_priv, new_state->shared_dpll, crtc, new_crtc_state);
> > +		verify_single_dpll_state(i915, new_state->shared_dpll, crtc, new_crtc_state);
> >   	if (old_state->shared_dpll &&
> >   	    old_state->shared_dpll != new_state->shared_dpll) {
> > @@ -13108,11 +13108,11 @@ intel_modeset_verify_crtc(struct drm_crtc *crtc,
> >   static void
> >   verify_disabled_dpll_state(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int i;
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++)
> > -		verify_single_dpll_state(dev_priv, &dev_priv->shared_dplls[i], NULL, NULL);
> > +	for (i = 0; i < i915->num_shared_dpll; i++)
> > +		verify_single_dpll_state(i915, &i915->shared_dplls[i], NULL, NULL);
> >   }
> >   static void
> > @@ -13127,7 +13127,7 @@ intel_modeset_verify_disabled(struct drm_device *dev,
> >   static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	/*
> >   	 * The scanline counter increments at the leading edge of hsync.
> > @@ -13156,7 +13156,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> >   	 * However if queried just before the start of vblank we'll get an
> >   	 * answer that's slightly in the future.
> >   	 */
> > -	if (IS_GEN(dev_priv, 2)) {
> > +	if (IS_GEN(i915, 2)) {
> >   		const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> >   		int vtotal;
> > @@ -13165,7 +13165,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> >   			vtotal /= 2;
> >   		crtc->scanline_offset = vtotal - 1;
> > -	} else if (HAS_DDI(dev_priv) &&
> > +	} else if (HAS_DDI(i915) &&
> >   		   intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) {
> >   		crtc->scanline_offset = 2;
> >   	} else
> > @@ -13174,12 +13174,12 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state)
> >   static void intel_modeset_clear_plls(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> >   	struct intel_crtc *crtc;
> >   	int i;
> > -	if (!dev_priv->display.crtc_compute_clock)
> > +	if (!i915->display.crtc_compute_clock)
> >   		return;
> >   	for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> > @@ -13311,7 +13311,7 @@ static int intel_modeset_all_pipes(struct drm_atomic_state *state)
> >   static int intel_modeset_checks(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> >   	struct intel_crtc *crtc;
> >   	int ret = 0, i;
> > @@ -13323,12 +13323,12 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   	/* keep the current setting */
> >   	if (!state->cdclk.force_min_cdclk_changed)
> > -		state->cdclk.force_min_cdclk = dev_priv->cdclk.force_min_cdclk;
> > +		state->cdclk.force_min_cdclk = i915->cdclk.force_min_cdclk;
> >   	state->modeset = true;
> > -	state->active_crtcs = dev_priv->active_crtcs;
> > -	state->cdclk.logical = dev_priv->cdclk.logical;
> > -	state->cdclk.actual = dev_priv->cdclk.actual;
> > +	state->active_crtcs = i915->active_crtcs;
> > +	state->cdclk.logical = i915->cdclk.logical;
> > +	state->cdclk.actual = i915->cdclk.actual;
> >   	state->cdclk.pipe = INVALID_PIPE;
> >   	for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> > @@ -13349,19 +13349,19 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   	 * mode set on this crtc.  For other crtcs we need to use the
> >   	 * adjusted_mode bits in the crtc directly.
> >   	 */
> > -	if (dev_priv->display.modeset_calc_cdclk) {
> > +	if (i915->display.modeset_calc_cdclk) {
> >   		enum pipe pipe;
> > -		ret = dev_priv->display.modeset_calc_cdclk(state);
> > +		ret = i915->display.modeset_calc_cdclk(state);
> >   		if (ret < 0)
> >   			return ret;
> >   		/*
> > -		 * Writes to dev_priv->cdclk.logical must protected by
> > +		 * Writes to i915->cdclk.logical must protected by
> >   		 * holding all the crtc locks, even if we don't end up
> >   		 * touching the hardware
> >   		 */
> > -		if (intel_cdclk_changed(&dev_priv->cdclk.logical,
> > +		if (intel_cdclk_changed(&i915->cdclk.logical,
> >   					&state->cdclk.logical)) {
> >   			ret = intel_lock_all_pipes(&state->base);
> >   			if (ret < 0)
> > @@ -13373,7 +13373,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   			struct drm_crtc_state *crtc_state;
> >   			pipe = ilog2(state->active_crtcs);
> > -			crtc = &intel_get_crtc_for_pipe(dev_priv, pipe)->base;
> > +			crtc = &intel_get_crtc_for_pipe(i915, pipe)->base;
> >   			crtc_state = drm_atomic_get_new_crtc_state(&state->base, crtc);
> >   			if (crtc_state && needs_modeset(crtc_state))
> >   				pipe = INVALID_PIPE;
> > @@ -13383,15 +13383,15 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   		/* All pipes must be switched off while we change the cdclk. */
> >   		if (pipe != INVALID_PIPE &&
> > -		    intel_cdclk_needs_cd2x_update(dev_priv,
> > -						  &dev_priv->cdclk.actual,
> > +		    intel_cdclk_needs_cd2x_update(i915,
> > +						  &i915->cdclk.actual,
> >   						  &state->cdclk.actual)) {
> >   			ret = intel_lock_all_pipes(&state->base);
> >   			if (ret < 0)
> >   				return ret;
> >   			state->cdclk.pipe = pipe;
> > -		} else if (intel_cdclk_needs_modeset(&dev_priv->cdclk.actual,
> > +		} else if (intel_cdclk_needs_modeset(&i915->cdclk.actual,
> >   						     &state->cdclk.actual)) {
> >   			ret = intel_modeset_all_pipes(&state->base);
> >   			if (ret < 0)
> > @@ -13410,7 +13410,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   	intel_modeset_clear_plls(state);
> > -	if (IS_HASWELL(dev_priv))
> > +	if (IS_HASWELL(i915))
> >   		return haswell_mode_set_planes_workaround(state);
> >   	return 0;
> > @@ -13424,11 +13424,11 @@ static int intel_modeset_checks(struct intel_atomic_state *state)
> >   static int calc_watermark_data(struct intel_atomic_state *state)
> >   {
> >   	struct drm_device *dev = state->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	/* Is there platform-specific watermark information to calculate? */
> > -	if (dev_priv->display.compute_global_watermarks)
> > -		return dev_priv->display.compute_global_watermarks(state);
> > +	if (i915->display.compute_global_watermarks)
> > +		return i915->display.compute_global_watermarks(state);
> >   	return 0;
> >   }
> > @@ -13441,7 +13441,7 @@ static int calc_watermark_data(struct intel_atomic_state *state)
> >   static int intel_atomic_check(struct drm_device *dev,
> >   			      struct drm_atomic_state *_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_atomic_state *state = to_intel_atomic_state(_state);
> >   	struct intel_crtc_state *old_crtc_state, *new_crtc_state;
> >   	struct intel_crtc *crtc;
> > @@ -13474,7 +13474,7 @@ static int intel_atomic_check(struct drm_device *dev,
> >   		if (ret)
> >   			goto fail;
> > -		if (intel_pipe_config_compare(dev_priv, old_crtc_state,
> > +		if (intel_pipe_config_compare(i915, old_crtc_state,
> >   					      new_crtc_state, true)) {
> >   			new_crtc_state->base.mode_changed = false;
> >   			new_crtc_state->update_pipe = true;
> > @@ -13493,7 +13493,7 @@ static int intel_atomic_check(struct drm_device *dev,
> >   		if (ret)
> >   			goto fail;
> >   	} else {
> > -		state->cdclk.logical = dev_priv->cdclk.logical;
> > +		state->cdclk.logical = i915->cdclk.logical;
> >   	}
> >   	ret = icl_add_linked_planes(state);
> > @@ -13504,7 +13504,7 @@ static int intel_atomic_check(struct drm_device *dev,
> >   	if (ret)
> >   		goto fail;
> > -	intel_fbc_choose_crtc(dev_priv, state);
> > +	intel_fbc_choose_crtc(i915, state);
> >   	ret = calc_watermark_data(state);
> >   	if (ret)
> >   		goto fail;
> > @@ -13564,7 +13564,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
> >   			      struct drm_crtc_state *new_crtc_state)
> >   {
> >   	struct drm_device *dev = crtc->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct intel_crtc_state *pipe_config = to_intel_crtc_state(new_crtc_state);
> >   	bool modeset = needs_modeset(new_crtc_state);
> > @@ -13574,7 +13574,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
> >   	if (modeset) {
> >   		update_scanline_offset(pipe_config);
> > -		dev_priv->display.crtc_enable(pipe_config, state);
> > +		i915->display.crtc_enable(pipe_config, state);
> >   		/* vblanks work again, re-enable pipe CRC. */
> >   		intel_crtc_enable_pipe_crc(intel_crtc);
> > @@ -13593,7 +13593,7 @@ static void intel_update_crtc(struct drm_crtc *crtc,
> >   	intel_begin_crtc_commit(to_intel_atomic_state(state), intel_crtc);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		skl_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc);
> >   	else
> >   		i9xx_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc);
> > @@ -13618,7 +13618,7 @@ static void intel_update_crtcs(struct drm_atomic_state *state)
> >   static void skl_update_crtcs(struct drm_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->dev);
> > +	struct drm_i915_private *i915 = to_i915(state->dev);
> >   	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> >   	struct drm_crtc *crtc;
> >   	struct intel_crtc *intel_crtc;
> > @@ -13628,7 +13628,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> >   	bool progress;
> >   	enum pipe pipe;
> >   	int i;
> > -	u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices;
> > +	u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices;
> >   	u8 required_slices = intel_state->wm_results.ddb.enabled_slices;
> >   	struct skl_ddb_entry entries[I915_MAX_PIPES] = {};
> > @@ -13638,8 +13638,8 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> >   			entries[i] = to_intel_crtc_state(old_crtc_state)->wm.skl.ddb;
> >   	/* If 2nd DBuf slice required, enable it here */
> > -	if (INTEL_GEN(dev_priv) >= 11 && required_slices > hw_enabled_slices)
> > -		icl_dbuf_slices_update(dev_priv, required_slices);
> > +	if (INTEL_GEN(i915) >= 11 && required_slices > hw_enabled_slices)
> > +		icl_dbuf_slices_update(i915, required_slices);
> >   	/*
> >   	 * Whenever the number of active pipes changes, we need to make sure we
> > @@ -13663,7 +13663,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> >   			if (skl_ddb_allocation_overlaps(&cstate->wm.skl.ddb,
> >   							entries,
> > -							INTEL_INFO(dev_priv)->num_pipes, i))
> > +							INTEL_INFO(i915)->num_pipes, i))
> >   				continue;
> >   			updated |= cmask;
> > @@ -13685,57 +13685,57 @@ static void skl_update_crtcs(struct drm_atomic_state *state)
> >   					  new_crtc_state);
> >   			if (vbl_wait)
> > -				intel_wait_for_vblank(dev_priv, pipe);
> > +				intel_wait_for_vblank(i915, pipe);
> >   			progress = true;
> >   		}
> >   	} while (progress);
> >   	/* If 2nd DBuf slice is no more required disable it */
> > -	if (INTEL_GEN(dev_priv) >= 11 && required_slices < hw_enabled_slices)
> > -		icl_dbuf_slices_update(dev_priv, required_slices);
> > +	if (INTEL_GEN(i915) >= 11 && required_slices < hw_enabled_slices)
> > +		icl_dbuf_slices_update(i915, required_slices);
> >   }
> > -static void intel_atomic_helper_free_state(struct drm_i915_private *dev_priv)
> > +static void intel_atomic_helper_free_state(struct drm_i915_private *i915)
> >   {
> >   	struct intel_atomic_state *state, *next;
> >   	struct llist_node *freed;
> > -	freed = llist_del_all(&dev_priv->atomic_helper.free_list);
> > +	freed = llist_del_all(&i915->atomic_helper.free_list);
> >   	llist_for_each_entry_safe(state, next, freed, freed)
> >   		drm_atomic_state_put(&state->base);
> >   }
> >   static void intel_atomic_helper_free_state_worker(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv), atomic_helper.free_work);
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915), atomic_helper.free_work);
> > -	intel_atomic_helper_free_state(dev_priv);
> > +	intel_atomic_helper_free_state(i915);
> >   }
> >   static void intel_atomic_commit_fence_wait(struct intel_atomic_state *intel_state)
> >   {
> >   	struct wait_queue_entry wait_fence, wait_reset;
> > -	struct drm_i915_private *dev_priv = to_i915(intel_state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_state->base.dev);
> >   	init_wait_entry(&wait_fence, 0);
> >   	init_wait_entry(&wait_reset, 0);
> >   	for (;;) {
> >   		prepare_to_wait(&intel_state->commit_ready.wait,
> >   				&wait_fence, TASK_UNINTERRUPTIBLE);
> > -		prepare_to_wait(&dev_priv->gpu_error.wait_queue,
> > +		prepare_to_wait(&i915->gpu_error.wait_queue,
> >   				&wait_reset, TASK_UNINTERRUPTIBLE);
> >   		if (i915_sw_fence_done(&intel_state->commit_ready)
> > -		    || test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags))
> > +		    || test_bit(I915_RESET_MODESET, &i915->gpu_error.flags))
> >   			break;
> >   		schedule();
> >   	}
> >   	finish_wait(&intel_state->commit_ready.wait, &wait_fence);
> > -	finish_wait(&dev_priv->gpu_error.wait_queue, &wait_reset);
> > +	finish_wait(&i915->gpu_error.wait_queue, &wait_reset);
> >   }
> >   static void intel_atomic_cleanup_work(struct work_struct *work)
> > @@ -13755,7 +13755,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   {
> >   	struct drm_device *dev = state->dev;
> >   	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_crtc_state *old_crtc_state, *new_crtc_state;
> >   	struct intel_crtc_state *new_intel_crtc_state, *old_intel_crtc_state;
> >   	struct drm_crtc *crtc;
> > @@ -13769,7 +13769,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   	drm_atomic_helper_wait_for_dependencies(state);
> >   	if (intel_state->modeset)
> > -		wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_MODESET);
> > +		wakeref = intel_display_power_get(i915, POWER_DOMAIN_MODESET);
> >   	for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) {
> >   		old_intel_crtc_state = to_intel_crtc_state(old_crtc_state);
> > @@ -13798,7 +13798,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   			 */
> >   			intel_crtc_disable_pipe_crc(intel_crtc);
> > -			dev_priv->display.crtc_disable(old_intel_crtc_state, state);
> > +			i915->display.crtc_disable(old_intel_crtc_state, state);
> >   			intel_crtc->active = false;
> >   			intel_fbc_disable(intel_crtc);
> >   			intel_disable_shared_dpll(old_intel_crtc_state);
> > @@ -13807,14 +13807,14 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   			 * Underruns don't always raise
> >   			 * interrupts, so check manually.
> >   			 */
> > -			intel_check_cpu_fifo_underruns(dev_priv);
> > -			intel_check_pch_fifo_underruns(dev_priv);
> > +			intel_check_cpu_fifo_underruns(i915);
> > +			intel_check_pch_fifo_underruns(i915);
> >   			/* FIXME unify this for all platforms */
> >   			if (!new_crtc_state->active &&
> > -			    !HAS_GMCH(dev_priv) &&
> > -			    dev_priv->display.initial_watermarks)
> > -				dev_priv->display.initial_watermarks(intel_state,
> > +			    !HAS_GMCH(i915) &&
> > +			    i915->display.initial_watermarks)
> > +				i915->display.initial_watermarks(intel_state,
> >   								     new_intel_crtc_state);
> >   		}
> >   	}
> > @@ -13826,9 +13826,9 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   	if (intel_state->modeset) {
> >   		drm_atomic_helper_update_legacy_modeset_state(state->dev, state);
> > -		intel_set_cdclk_pre_plane_update(dev_priv,
> > +		intel_set_cdclk_pre_plane_update(i915,
> >   						 &intel_state->cdclk.actual,
> > -						 &dev_priv->cdclk.actual,
> > +						 &i915->cdclk.actual,
> >   						 intel_state->cdclk.pipe);
> >   		/*
> > @@ -13836,7 +13836,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   		 * have more then one pipe enabled
> >   		 */
> >   		if (!intel_can_enable_sagv(state))
> > -			intel_disable_sagv(dev_priv);
> > +			intel_disable_sagv(i915);
> >   		intel_modeset_verify_disabled(dev, state);
> >   	}
> > @@ -13856,12 +13856,12 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   	}
> >   	/* Now enable the clocks, plane, pipe, and connectors that we set up. */
> > -	dev_priv->display.update_crtcs(state);
> > +	i915->display.update_crtcs(state);
> >   	if (intel_state->modeset)
> > -		intel_set_cdclk_post_plane_update(dev_priv,
> > +		intel_set_cdclk_post_plane_update(i915,
> >   						  &intel_state->cdclk.actual,
> > -						  &dev_priv->cdclk.actual,
> > +						  &i915->cdclk.actual,
> >   						  intel_state->cdclk.pipe);
> >   	/* FIXME: We should call drm_atomic_helper_commit_hw_done() here
> > @@ -13895,8 +13895,8 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   	for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) {
> >   		new_intel_crtc_state = to_intel_crtc_state(new_crtc_state);
> > -		if (dev_priv->display.optimize_watermarks)
> > -			dev_priv->display.optimize_watermarks(intel_state,
> > +		if (i915->display.optimize_watermarks)
> > +			i915->display.optimize_watermarks(intel_state,
> >   							      new_intel_crtc_state);
> >   	}
> > @@ -13904,7 +13904,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   		intel_post_plane_update(to_intel_crtc_state(old_crtc_state));
> >   		if (put_domains[i])
> > -			modeset_put_power_domains(dev_priv, put_domains[i]);
> > +			modeset_put_power_domains(i915, put_domains[i]);
> >   		intel_modeset_verify_crtc(crtc, state, old_crtc_state, new_crtc_state);
> >   	}
> > @@ -13913,7 +13913,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   		intel_verify_planes(intel_state);
> >   	if (intel_state->modeset && intel_can_enable_sagv(state))
> > -		intel_enable_sagv(dev_priv);
> > +		intel_enable_sagv(i915);
> >   	drm_atomic_helper_commit_hw_done(state);
> > @@ -13924,10 +13924,10 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state)
> >   		 * so enable debugging for the next modeset - and hope we catch
> >   		 * the culprit.
> >   		 */
> > -		intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore);
> > -		intel_display_power_put(dev_priv, POWER_DOMAIN_MODESET, wakeref);
> > +		intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore);
> > +		intel_display_power_put(i915, POWER_DOMAIN_MODESET, wakeref);
> >   	}
> > -	intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> > +	intel_runtime_pm_put(i915, intel_state->wakeref);
> >   	/*
> >   	 * Defer the cleanup of the old state to a separate worker to not
> > @@ -14003,10 +14003,10 @@ static int intel_atomic_commit(struct drm_device *dev,
> >   			       bool nonblock)
> >   {
> >   	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int ret = 0;
> > -	intel_state->wakeref = intel_runtime_pm_get(dev_priv);
> > +	intel_state->wakeref = intel_runtime_pm_get(i915);
> >   	drm_atomic_state_get(state);
> >   	i915_sw_fence_init(&intel_state->commit_ready,
> > @@ -14029,7 +14029,7 @@ static int intel_atomic_commit(struct drm_device *dev,
> >   	 * FIXME doing watermarks and fb cleanup from a vblank worker
> >   	 * (assuming we had any) would solve these problems.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 9 && state->legacy_cursor_update) {
> > +	if (INTEL_GEN(i915) < 9 && state->legacy_cursor_update) {
> >   		struct intel_crtc_state *new_crtc_state;
> >   		struct intel_crtc *crtc;
> >   		int i;
> > @@ -14044,7 +14044,7 @@ static int intel_atomic_commit(struct drm_device *dev,
> >   	if (ret) {
> >   		DRM_DEBUG_ATOMIC("Preparing state failed with %i\n", ret);
> >   		i915_sw_fence_commit(&intel_state->commit_ready);
> > -		intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> > +		intel_runtime_pm_put(i915, intel_state->wakeref);
> >   		return ret;
> >   	}
> > @@ -14056,21 +14056,21 @@ static int intel_atomic_commit(struct drm_device *dev,
> >   		i915_sw_fence_commit(&intel_state->commit_ready);
> >   		drm_atomic_helper_cleanup_planes(dev, state);
> > -		intel_runtime_pm_put(dev_priv, intel_state->wakeref);
> > +		intel_runtime_pm_put(i915, intel_state->wakeref);
> >   		return ret;
> >   	}
> > -	dev_priv->wm.distrust_bios_wm = false;
> > +	i915->wm.distrust_bios_wm = false;
> >   	intel_shared_dpll_swap_state(state);
> >   	intel_atomic_track_fbs(state);
> >   	if (intel_state->modeset) {
> > -		memcpy(dev_priv->min_cdclk, intel_state->min_cdclk,
> > +		memcpy(i915->min_cdclk, intel_state->min_cdclk,
> >   		       sizeof(intel_state->min_cdclk));
> > -		memcpy(dev_priv->min_voltage_level,
> > +		memcpy(i915->min_voltage_level,
> >   		       intel_state->min_voltage_level,
> >   		       sizeof(intel_state->min_voltage_level));
> > -		dev_priv->active_crtcs = intel_state->active_crtcs;
> > -		dev_priv->cdclk.force_min_cdclk =
> > +		i915->active_crtcs = intel_state->active_crtcs;
> > +		i915->cdclk.force_min_cdclk =
> >   			intel_state->cdclk.force_min_cdclk;
> >   		intel_cdclk_swap_state(intel_state);
> > @@ -14081,12 +14081,12 @@ static int intel_atomic_commit(struct drm_device *dev,
> >   	i915_sw_fence_commit(&intel_state->commit_ready);
> >   	if (nonblock && intel_state->modeset) {
> > -		queue_work(dev_priv->modeset_wq, &state->commit_work);
> > +		queue_work(i915->modeset_wq, &state->commit_work);
> >   	} else if (nonblock) {
> >   		queue_work(system_unbound_wq, &state->commit_work);
> >   	} else {
> >   		if (intel_state->modeset)
> > -			flush_workqueue(dev_priv->modeset_wq);
> > +			flush_workqueue(i915->modeset_wq);
> >   		intel_atomic_commit_tail(state);
> >   	}
> > @@ -14166,14 +14166,14 @@ static void add_rps_boost_after_vblank(struct drm_crtc *crtc,
> >   static int intel_plane_pin_fb(struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	struct drm_framebuffer *fb = plane_state->base.fb;
> >   	struct i915_vma *vma;
> >   	if (plane->id == PLANE_CURSOR &&
> > -	    INTEL_INFO(dev_priv)->display.cursor_needs_physical) {
> > +	    INTEL_INFO(i915)->display.cursor_needs_physical) {
> >   		struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> > -		const int align = intel_cursor_alignment(dev_priv);
> > +		const int align = intel_cursor_alignment(i915);
> >   		int err;
> >   		err = i915_gem_object_attach_phys(obj, align);
> > @@ -14231,7 +14231,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> >   {
> >   	struct intel_atomic_state *intel_state =
> >   		to_intel_atomic_state(new_state->state);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->dev);
> >   	struct drm_framebuffer *fb = new_state->fb;
> >   	struct drm_i915_gem_object *obj = intel_fb_obj(fb);
> >   	struct drm_i915_gem_object *old_obj = intel_fb_obj(plane->state->fb);
> > @@ -14279,7 +14279,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> >   	if (ret)
> >   		return ret;
> > -	ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex);
> > +	ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
> >   	if (ret) {
> >   		i915_gem_object_unpin_pages(obj);
> >   		return ret;
> > @@ -14287,7 +14287,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> >   	ret = intel_plane_pin_fb(to_intel_plane_state(new_state));
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	i915_gem_object_unpin_pages(obj);
> >   	if (ret)
> >   		return ret;
> > @@ -14323,7 +14323,7 @@ intel_prepare_plane_fb(struct drm_plane *plane,
> >   	 * maximum clocks following a vblank miss (see do_rps_boost()).
> >   	 */
> >   	if (!intel_state->rps_interactive) {
> > -		intel_rps_mark_interactive(dev_priv, true);
> > +		intel_rps_mark_interactive(i915, true);
> >   		intel_state->rps_interactive = true;
> >   	}
> > @@ -14345,17 +14345,17 @@ intel_cleanup_plane_fb(struct drm_plane *plane,
> >   {
> >   	struct intel_atomic_state *intel_state =
> >   		to_intel_atomic_state(old_state->state);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->dev);
> >   	if (intel_state->rps_interactive) {
> > -		intel_rps_mark_interactive(dev_priv, false);
> > +		intel_rps_mark_interactive(i915, false);
> >   		intel_state->rps_interactive = false;
> >   	}
> >   	/* Should only be called after a successful intel_prepare_plane_fb()! */
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	intel_plane_unpin_fb(to_intel_plane_state(old_state));
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   }
> >   int
> > @@ -14363,7 +14363,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> >   	      u32 pixel_format)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int max_scale, mult;
> >   	int crtc_clock, max_dotclk, tmpclk1, tmpclk2;
> > @@ -14373,7 +14373,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> >   	crtc_clock = crtc_state->base.adjusted_mode.crtc_clock;
> >   	max_dotclk = to_intel_atomic_state(crtc_state->base.state)->cdclk.logical.cdclk;
> > -	if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> > +	if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10)
> >   		max_dotclk *= 2;
> >   	if (WARN_ON_ONCE(!crtc_clock || max_dotclk < crtc_clock))
> > @@ -14396,7 +14396,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state,
> >   static void intel_begin_crtc_commit(struct intel_atomic_state *state,
> >   				    struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_crtc_state *old_crtc_state =
> >   		intel_atomic_get_old_crtc_state(state, crtc);
> >   	struct intel_crtc_state *new_crtc_state =
> > @@ -14415,31 +14415,31 @@ static void intel_begin_crtc_commit(struct intel_atomic_state *state,
> >   	if (new_crtc_state->update_pipe)
> >   		intel_update_pipe_config(old_crtc_state, new_crtc_state);
> > -	else if (INTEL_GEN(dev_priv) >= 9)
> > +	else if (INTEL_GEN(i915) >= 9)
> >   		skl_detach_scalers(new_crtc_state);
> > -	if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv))
> > +	if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915))
> >   		bdw_set_pipemisc(new_crtc_state);
> >   out:
> > -	if (dev_priv->display.atomic_update_watermarks)
> > -		dev_priv->display.atomic_update_watermarks(state,
> > +	if (i915->display.atomic_update_watermarks)
> > +		i915->display.atomic_update_watermarks(state,
> >   							   new_crtc_state);
> >   }
> >   void intel_crtc_arm_fifo_underrun(struct intel_crtc *crtc,
> >   				  struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	if (!IS_GEN(dev_priv, 2))
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true);
> > +	if (!IS_GEN(i915, 2))
> > +		intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true);
> >   	if (crtc_state->has_pch_encoder) {
> >   		enum pipe pch_transcoder =
> >   			intel_crtc_pch_transcoder(crtc);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder, true);
> > +		intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder, true);
> >   	}
> >   }
> > @@ -14559,7 +14559,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> >   			   u32 src_w, u32 src_h,
> >   			   struct drm_modeset_acquire_ctx *ctx)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	int ret;
> >   	struct drm_plane_state *old_plane_state, *new_plane_state;
> >   	struct intel_plane *intel_plane = to_intel_plane(plane);
> > @@ -14626,7 +14626,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> >   	if (ret)
> >   		goto out_free;
> > -	ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex);
> > +	ret = mutex_lock_interruptible(&i915->drm.struct_mutex);
> >   	if (ret)
> >   		goto out_free;
> > @@ -14664,7 +14664,7 @@ intel_legacy_cursor_update(struct drm_plane *plane,
> >   	intel_plane_unpin_fb(to_intel_plane_state(old_plane_state));
> >   out_unlock:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   out_free:
> >   	if (new_crtc_state)
> >   		intel_crtc_destroy_state(crtc, &new_crtc_state->base);
> > @@ -14691,25 +14691,25 @@ static const struct drm_plane_funcs intel_cursor_plane_funcs = {
> >   	.format_mod_supported = intel_cursor_format_mod_supported,
> >   };
> > -static bool i9xx_plane_has_fbc(struct drm_i915_private *dev_priv,
> > +static bool i9xx_plane_has_fbc(struct drm_i915_private *i915,
> >   			       enum i9xx_plane_id i9xx_plane)
> >   {
> > -	if (!HAS_FBC(dev_priv))
> > +	if (!HAS_FBC(i915))
> >   		return false;
> > -	if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> > +	if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> >   		return i9xx_plane == PLANE_A; /* tied to pipe A */
> > -	else if (IS_IVYBRIDGE(dev_priv))
> > +	else if (IS_IVYBRIDGE(i915))
> >   		return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B ||
> >   			i9xx_plane == PLANE_C;
> > -	else if (INTEL_GEN(dev_priv) >= 4)
> > +	else if (INTEL_GEN(i915) >= 4)
> >   		return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B;
> >   	else
> >   		return i9xx_plane == PLANE_A;
> >   }
> >   static struct intel_plane *
> > -intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +intel_primary_plane_create(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	struct intel_plane *plane;
> >   	const struct drm_plane_funcs *plane_funcs;
> > @@ -14720,8 +14720,8 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	int num_formats;
> >   	int ret;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		return skl_universal_plane_create(dev_priv, pipe,
> > +	if (INTEL_GEN(i915) >= 9)
> > +		return skl_universal_plane_create(i915, pipe,
> >   						  PLANE_PRIMARY);
> >   	plane = intel_plane_alloc();
> > @@ -14733,21 +14733,21 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	 * On gen2/3 only plane A can do FBC, but the panel fitter and LVDS
> >   	 * port is hooked to pipe B. Hence we want plane A feeding pipe B.
> >   	 */
> > -	if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) < 4)
> > +	if (HAS_FBC(i915) && INTEL_GEN(i915) < 4)
> >   		plane->i9xx_plane = (enum i9xx_plane_id) !pipe;
> >   	else
> >   		plane->i9xx_plane = (enum i9xx_plane_id) pipe;
> >   	plane->id = PLANE_PRIMARY;
> >   	plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane->id);
> > -	plane->has_fbc = i9xx_plane_has_fbc(dev_priv, plane->i9xx_plane);
> > +	plane->has_fbc = i9xx_plane_has_fbc(i915, plane->i9xx_plane);
> >   	if (plane->has_fbc) {
> > -		struct intel_fbc *fbc = &dev_priv->fbc;
> > +		struct intel_fbc *fbc = &i915->fbc;
> >   		fbc->possible_framebuffer_bits |= plane->frontbuffer_bit;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		formats = i965_primary_formats;
> >   		num_formats = ARRAY_SIZE(i965_primary_formats);
> >   		modifiers = i9xx_format_modifiers;
> > @@ -14775,14 +14775,14 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	possible_crtcs = BIT(pipe);
> > -	if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv))
> > -		ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> > +	if (INTEL_GEN(i915) >= 5 || IS_G4X(i915))
> > +		ret = drm_universal_plane_init(&i915->drm, &plane->base,
> >   					       possible_crtcs, plane_funcs,
> >   					       formats, num_formats, modifiers,
> >   					       DRM_PLANE_TYPE_PRIMARY,
> >   					       "primary %c", pipe_name(pipe));
> >   	else
> > -		ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> > +		ret = drm_universal_plane_init(&i915->drm, &plane->base,
> >   					       possible_crtcs, plane_funcs,
> >   					       formats, num_formats, modifiers,
> >   					       DRM_PLANE_TYPE_PRIMARY,
> > @@ -14791,18 +14791,18 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	if (ret)
> >   		goto fail;
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> >   		supported_rotations =
> >   			DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 |
> >   			DRM_MODE_REFLECT_X;
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > +	} else if (INTEL_GEN(i915) >= 4) {
> >   		supported_rotations =
> >   			DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180;
> >   	} else {
> >   		supported_rotations = DRM_MODE_ROTATE_0;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		drm_plane_create_rotation_property(&plane->base,
> >   						   DRM_MODE_ROTATE_0,
> >   						   supported_rotations);
> > @@ -14818,7 +14818,7 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   }
> >   static struct intel_plane *
> > -intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> > +intel_cursor_plane_create(struct drm_i915_private *i915,
> >   			  enum pipe pipe)
> >   {
> >   	unsigned int possible_crtcs;
> > @@ -14834,7 +14834,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> >   	cursor->id = PLANE_CURSOR;
> >   	cursor->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, cursor->id);
> > -	if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) {
> > +	if (IS_I845G(i915) || IS_I865G(i915)) {
> >   		cursor->max_stride = i845_cursor_max_stride;
> >   		cursor->update_plane = i845_update_cursor;
> >   		cursor->disable_plane = i845_disable_cursor;
> > @@ -14851,12 +14851,12 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> >   	cursor->cursor.base = ~0;
> >   	cursor->cursor.cntl = ~0;
> > -	if (IS_I845G(dev_priv) || IS_I865G(dev_priv) || HAS_CUR_FBC(dev_priv))
> > +	if (IS_I845G(i915) || IS_I865G(i915) || HAS_CUR_FBC(i915))
> >   		cursor->cursor.size = ~0;
> >   	possible_crtcs = BIT(pipe);
> > -	ret = drm_universal_plane_init(&dev_priv->drm, &cursor->base,
> > +	ret = drm_universal_plane_init(&i915->drm, &cursor->base,
> >   				       possible_crtcs, &intel_cursor_plane_funcs,
> >   				       intel_cursor_formats,
> >   				       ARRAY_SIZE(intel_cursor_formats),
> > @@ -14866,7 +14866,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv,
> >   	if (ret)
> >   		goto fail;
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		drm_plane_create_rotation_property(&cursor->base,
> >   						   DRM_MODE_ROTATE_0,
> >   						   DRM_MODE_ROTATE_0 |
> > @@ -14887,10 +14887,10 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc,
> >   {
> >   	struct intel_crtc_scaler_state *scaler_state =
> >   		&crtc_state->scaler_state;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	int i;
> > -	crtc->num_scalers = RUNTIME_INFO(dev_priv)->num_scalers[crtc->pipe];
> > +	crtc->num_scalers = RUNTIME_INFO(i915)->num_scalers[crtc->pipe];
> >   	if (!crtc->num_scalers)
> >   		return;
> > @@ -14904,7 +14904,7 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc,
> >   	scaler_state->scaler_id = -1;
> >   }
> > -static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +static int intel_crtc_init(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	struct intel_crtc *intel_crtc;
> >   	struct intel_crtc_state *crtc_state = NULL;
> > @@ -14924,17 +14924,17 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	__drm_atomic_helper_crtc_reset(&intel_crtc->base, &crtc_state->base);
> >   	intel_crtc->config = crtc_state;
> > -	primary = intel_primary_plane_create(dev_priv, pipe);
> > +	primary = intel_primary_plane_create(i915, pipe);
> >   	if (IS_ERR(primary)) {
> >   		ret = PTR_ERR(primary);
> >   		goto fail;
> >   	}
> >   	intel_crtc->plane_ids_mask |= BIT(primary->id);
> > -	for_each_sprite(dev_priv, pipe, sprite) {
> > +	for_each_sprite(i915, pipe, sprite) {
> >   		struct intel_plane *plane;
> > -		plane = intel_sprite_plane_create(dev_priv, pipe, sprite);
> > +		plane = intel_sprite_plane_create(i915, pipe, sprite);
> >   		if (IS_ERR(plane)) {
> >   			ret = PTR_ERR(plane);
> >   			goto fail;
> > @@ -14942,14 +14942,14 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   		intel_crtc->plane_ids_mask |= BIT(plane->id);
> >   	}
> > -	cursor = intel_cursor_plane_create(dev_priv, pipe);
> > +	cursor = intel_cursor_plane_create(i915, pipe);
> >   	if (IS_ERR(cursor)) {
> >   		ret = PTR_ERR(cursor);
> >   		goto fail;
> >   	}
> >   	intel_crtc->plane_ids_mask |= BIT(cursor->id);
> > -	ret = drm_crtc_init_with_planes(&dev_priv->drm, &intel_crtc->base,
> > +	ret = drm_crtc_init_with_planes(&i915->drm, &intel_crtc->base,
> >   					&primary->base, &cursor->base,
> >   					&intel_crtc_funcs,
> >   					"pipe %c", pipe_name(pipe));
> > @@ -14961,16 +14961,16 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	/* initialize shared scalers */
> >   	intel_crtc_init_scalers(intel_crtc, crtc_state);
> > -	BUG_ON(pipe >= ARRAY_SIZE(dev_priv->pipe_to_crtc_mapping) ||
> > -	       dev_priv->pipe_to_crtc_mapping[pipe] != NULL);
> > -	dev_priv->pipe_to_crtc_mapping[pipe] = intel_crtc;
> > +	BUG_ON(pipe >= ARRAY_SIZE(i915->pipe_to_crtc_mapping) ||
> > +	       i915->pipe_to_crtc_mapping[pipe] != NULL);
> > +	i915->pipe_to_crtc_mapping[pipe] = intel_crtc;
> > -	if (INTEL_GEN(dev_priv) < 9) {
> > +	if (INTEL_GEN(i915) < 9) {
> >   		enum i9xx_plane_id i9xx_plane = primary->i9xx_plane;
> > -		BUG_ON(i9xx_plane >= ARRAY_SIZE(dev_priv->plane_to_crtc_mapping) ||
> > -		       dev_priv->plane_to_crtc_mapping[i9xx_plane] != NULL);
> > -		dev_priv->plane_to_crtc_mapping[i9xx_plane] = intel_crtc;
> > +		BUG_ON(i9xx_plane >= ARRAY_SIZE(i915->plane_to_crtc_mapping) ||
> > +		       i915->plane_to_crtc_mapping[i9xx_plane] != NULL);
> > +		i915->plane_to_crtc_mapping[i9xx_plane] = intel_crtc;
> >   	}
> >   	drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs);
> > @@ -15026,29 +15026,29 @@ static int intel_encoder_clones(struct intel_encoder *encoder)
> >   	return index_mask;
> >   }
> > -static bool ilk_has_edp_a(struct drm_i915_private *dev_priv)
> > +static bool ilk_has_edp_a(struct drm_i915_private *i915)
> >   {
> > -	if (!IS_MOBILE(dev_priv))
> > +	if (!IS_MOBILE(i915))
> >   		return false;
> >   	if ((I915_READ(DP_A) & DP_DETECTED) == 0)
> >   		return false;
> > -	if (IS_GEN(dev_priv, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE))
> > +	if (IS_GEN(i915, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE))
> >   		return false;
> >   	return true;
> >   }
> > -static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv)
> > +static bool intel_ddi_crt_present(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return false;
> > -	if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv))
> > +	if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915))
> >   		return false;
> > -	if (HAS_PCH_LPT_H(dev_priv) &&
> > +	if (HAS_PCH_LPT_H(i915) &&
> >   	    I915_READ(SFUSE_STRAP) & SFUSE_STRAP_CRT_DISABLED)
> >   		return false;
> > @@ -15056,24 +15056,24 @@ static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv)
> >   	if (I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_A_4_LANES)
> >   		return false;
> > -	if (!dev_priv->vbt.int_crt_support)
> > +	if (!i915->vbt.int_crt_support)
> >   		return false;
> >   	return true;
> >   }
> > -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv)
> > +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915)
> >   {
> >   	int pps_num;
> >   	int pps_idx;
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		return;
> >   	/*
> >   	 * This w/a is needed at least on CPT/PPT, but to be sure apply it
> >   	 * everywhere where registers can be write protected.
> >   	 */
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		pps_num = 2;
> >   	else
> >   		pps_num = 1;
> > @@ -15086,65 +15086,65 @@ void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void intel_pps_init(struct drm_i915_private *dev_priv)
> > +static void intel_pps_init(struct drm_i915_private *i915)
> >   {
> > -	if (HAS_PCH_SPLIT(dev_priv) || IS_GEN9_LP(dev_priv))
> > -		dev_priv->pps_mmio_base = PCH_PPS_BASE;
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->pps_mmio_base = VLV_PPS_BASE;
> > +	if (HAS_PCH_SPLIT(i915) || IS_GEN9_LP(i915))
> > +		i915->pps_mmio_base = PCH_PPS_BASE;
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->pps_mmio_base = VLV_PPS_BASE;
> >   	else
> > -		dev_priv->pps_mmio_base = PPS_BASE;
> > +		i915->pps_mmio_base = PPS_BASE;
> > -	intel_pps_unlock_regs_wa(dev_priv);
> > +	intel_pps_unlock_regs_wa(i915);
> >   }
> > -static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> > +static void intel_setup_outputs(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> >   	bool dpd_is_edp = false;
> > -	intel_pps_init(dev_priv);
> > +	intel_pps_init(i915);
> > -	if (!HAS_DISPLAY(dev_priv))
> > +	if (!HAS_DISPLAY(i915))
> >   		return;
> > -	if (IS_ELKHARTLAKE(dev_priv)) {
> > -		intel_ddi_init(dev_priv, PORT_A);
> > -		intel_ddi_init(dev_priv, PORT_B);
> > -		intel_ddi_init(dev_priv, PORT_C);
> > -		icl_dsi_init(dev_priv);
> > -	} else if (INTEL_GEN(dev_priv) >= 11) {
> > -		intel_ddi_init(dev_priv, PORT_A);
> > -		intel_ddi_init(dev_priv, PORT_B);
> > -		intel_ddi_init(dev_priv, PORT_C);
> > -		intel_ddi_init(dev_priv, PORT_D);
> > -		intel_ddi_init(dev_priv, PORT_E);
> > +	if (IS_ELKHARTLAKE(i915)) {
> > +		intel_ddi_init(i915, PORT_A);
> > +		intel_ddi_init(i915, PORT_B);
> > +		intel_ddi_init(i915, PORT_C);
> > +		icl_dsi_init(i915);
> > +	} else if (INTEL_GEN(i915) >= 11) {
> > +		intel_ddi_init(i915, PORT_A);
> > +		intel_ddi_init(i915, PORT_B);
> > +		intel_ddi_init(i915, PORT_C);
> > +		intel_ddi_init(i915, PORT_D);
> > +		intel_ddi_init(i915, PORT_E);
> >   		/*
> >   		 * On some ICL SKUs port F is not present. No strap bits for
> >   		 * this, so rely on VBT.
> >   		 * Work around broken VBTs on SKUs known to have no port F.
> >   		 */
> > -		if (IS_ICL_WITH_PORT_F(dev_priv) &&
> > -		    intel_bios_is_port_present(dev_priv, PORT_F))
> > -			intel_ddi_init(dev_priv, PORT_F);
> > +		if (IS_ICL_WITH_PORT_F(i915) &&
> > +		    intel_bios_is_port_present(i915, PORT_F))
> > +			intel_ddi_init(i915, PORT_F);
> > -		icl_dsi_init(dev_priv);
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > +		icl_dsi_init(i915);
> > +	} else if (IS_GEN9_LP(i915)) {
> >   		/*
> >   		 * FIXME: Broxton doesn't support port detection via the
> >   		 * DDI_BUF_CTL_A or SFUSE_STRAP registers, find another way to
> >   		 * detect the ports.
> >   		 */
> > -		intel_ddi_init(dev_priv, PORT_A);
> > -		intel_ddi_init(dev_priv, PORT_B);
> > -		intel_ddi_init(dev_priv, PORT_C);
> > +		intel_ddi_init(i915, PORT_A);
> > +		intel_ddi_init(i915, PORT_B);
> > +		intel_ddi_init(i915, PORT_C);
> > -		vlv_dsi_init(dev_priv);
> > -	} else if (HAS_DDI(dev_priv)) {
> > +		vlv_dsi_init(i915);
> > +	} else if (HAS_DDI(i915)) {
> >   		int found;
> > -		if (intel_ddi_crt_present(dev_priv))
> > -			intel_crt_init(dev_priv);
> > +		if (intel_ddi_crt_present(i915))
> > +			intel_crt_init(i915);
> >   		/*
> >   		 * Haswell uses DDI functions to detect digital outputs.
> > @@ -15153,29 +15153,29 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> >   		 */
> >   		found = I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_INIT_DISPLAY_DETECTED;
> >   		/* WaIgnoreDDIAStrap: skl */
> > -		if (found || IS_GEN9_BC(dev_priv))
> > -			intel_ddi_init(dev_priv, PORT_A);
> > +		if (found || IS_GEN9_BC(i915))
> > +			intel_ddi_init(i915, PORT_A);
> >   		/* DDI B, C, D, and F detection is indicated by the SFUSE_STRAP
> >   		 * register */
> >   		found = I915_READ(SFUSE_STRAP);
> >   		if (found & SFUSE_STRAP_DDIB_DETECTED)
> > -			intel_ddi_init(dev_priv, PORT_B);
> > +			intel_ddi_init(i915, PORT_B);
> >   		if (found & SFUSE_STRAP_DDIC_DETECTED)
> > -			intel_ddi_init(dev_priv, PORT_C);
> > +			intel_ddi_init(i915, PORT_C);
> >   		if (found & SFUSE_STRAP_DDID_DETECTED)
> > -			intel_ddi_init(dev_priv, PORT_D);
> > +			intel_ddi_init(i915, PORT_D);
> >   		if (found & SFUSE_STRAP_DDIF_DETECTED)
> > -			intel_ddi_init(dev_priv, PORT_F);
> > +			intel_ddi_init(i915, PORT_F);
> >   		/*
> >   		 * On SKL we don't have a way to detect DDI-E so we rely on VBT.
> >   		 */
> > -		if (IS_GEN9_BC(dev_priv) &&
> > -		    intel_bios_is_port_present(dev_priv, PORT_E))
> > -			intel_ddi_init(dev_priv, PORT_E);
> > +		if (IS_GEN9_BC(i915) &&
> > +		    intel_bios_is_port_present(i915, PORT_E))
> > +			intel_ddi_init(i915, PORT_E);
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		int found;
> >   		/*
> > @@ -15183,39 +15183,39 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> >   		 * to prevent the registration of both eDP and LVDS and the
> >   		 * incorrect sharing of the PPS.
> >   		 */
> > -		intel_lvds_init(dev_priv);
> > -		intel_crt_init(dev_priv);
> > +		intel_lvds_init(i915);
> > +		intel_crt_init(i915);
> > -		dpd_is_edp = intel_dp_is_port_edp(dev_priv, PORT_D);
> > +		dpd_is_edp = intel_dp_is_port_edp(i915, PORT_D);
> > -		if (ilk_has_edp_a(dev_priv))
> > -			intel_dp_init(dev_priv, DP_A, PORT_A);
> > +		if (ilk_has_edp_a(i915))
> > +			intel_dp_init(i915, DP_A, PORT_A);
> >   		if (I915_READ(PCH_HDMIB) & SDVO_DETECTED) {
> >   			/* PCH SDVOB multiplex with HDMIB */
> > -			found = intel_sdvo_init(dev_priv, PCH_SDVOB, PORT_B);
> > +			found = intel_sdvo_init(i915, PCH_SDVOB, PORT_B);
> >   			if (!found)
> > -				intel_hdmi_init(dev_priv, PCH_HDMIB, PORT_B);
> > +				intel_hdmi_init(i915, PCH_HDMIB, PORT_B);
> >   			if (!found && (I915_READ(PCH_DP_B) & DP_DETECTED))
> > -				intel_dp_init(dev_priv, PCH_DP_B, PORT_B);
> > +				intel_dp_init(i915, PCH_DP_B, PORT_B);
> >   		}
> >   		if (I915_READ(PCH_HDMIC) & SDVO_DETECTED)
> > -			intel_hdmi_init(dev_priv, PCH_HDMIC, PORT_C);
> > +			intel_hdmi_init(i915, PCH_HDMIC, PORT_C);
> >   		if (!dpd_is_edp && I915_READ(PCH_HDMID) & SDVO_DETECTED)
> > -			intel_hdmi_init(dev_priv, PCH_HDMID, PORT_D);
> > +			intel_hdmi_init(i915, PCH_HDMID, PORT_D);
> >   		if (I915_READ(PCH_DP_C) & DP_DETECTED)
> > -			intel_dp_init(dev_priv, PCH_DP_C, PORT_C);
> > +			intel_dp_init(i915, PCH_DP_C, PORT_C);
> >   		if (I915_READ(PCH_DP_D) & DP_DETECTED)
> > -			intel_dp_init(dev_priv, PCH_DP_D, PORT_D);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +			intel_dp_init(i915, PCH_DP_D, PORT_D);
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		bool has_edp, has_port;
> > -		if (IS_VALLEYVIEW(dev_priv) && dev_priv->vbt.int_crt_support)
> > -			intel_crt_init(dev_priv);
> > +		if (IS_VALLEYVIEW(i915) && i915->vbt.int_crt_support)
> > +			intel_crt_init(i915);
> >   		/*
> >   		 * The DP_DETECTED bit is the latched state of the DDC
> > @@ -15232,97 +15232,97 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv)
> >   		 * trust the port type the VBT declares as we've seen at least
> >   		 * HDMI ports that the VBT claim are DP or eDP.
> >   		 */
> > -		has_edp = intel_dp_is_port_edp(dev_priv, PORT_B);
> > -		has_port = intel_bios_is_port_present(dev_priv, PORT_B);
> > +		has_edp = intel_dp_is_port_edp(i915, PORT_B);
> > +		has_port = intel_bios_is_port_present(i915, PORT_B);
> >   		if (I915_READ(VLV_DP_B) & DP_DETECTED || has_port)
> > -			has_edp &= intel_dp_init(dev_priv, VLV_DP_B, PORT_B);
> > +			has_edp &= intel_dp_init(i915, VLV_DP_B, PORT_B);
> >   		if ((I915_READ(VLV_HDMIB) & SDVO_DETECTED || has_port) && !has_edp)
> > -			intel_hdmi_init(dev_priv, VLV_HDMIB, PORT_B);
> > +			intel_hdmi_init(i915, VLV_HDMIB, PORT_B);
> > -		has_edp = intel_dp_is_port_edp(dev_priv, PORT_C);
> > -		has_port = intel_bios_is_port_present(dev_priv, PORT_C);
> > +		has_edp = intel_dp_is_port_edp(i915, PORT_C);
> > +		has_port = intel_bios_is_port_present(i915, PORT_C);
> >   		if (I915_READ(VLV_DP_C) & DP_DETECTED || has_port)
> > -			has_edp &= intel_dp_init(dev_priv, VLV_DP_C, PORT_C);
> > +			has_edp &= intel_dp_init(i915, VLV_DP_C, PORT_C);
> >   		if ((I915_READ(VLV_HDMIC) & SDVO_DETECTED || has_port) && !has_edp)
> > -			intel_hdmi_init(dev_priv, VLV_HDMIC, PORT_C);
> > +			intel_hdmi_init(i915, VLV_HDMIC, PORT_C);
> > -		if (IS_CHERRYVIEW(dev_priv)) {
> > +		if (IS_CHERRYVIEW(i915)) {
> >   			/*
> >   			 * eDP not supported on port D,
> >   			 * so no need to worry about it
> >   			 */
> > -			has_port = intel_bios_is_port_present(dev_priv, PORT_D);
> > +			has_port = intel_bios_is_port_present(i915, PORT_D);
> >   			if (I915_READ(CHV_DP_D) & DP_DETECTED || has_port)
> > -				intel_dp_init(dev_priv, CHV_DP_D, PORT_D);
> > +				intel_dp_init(i915, CHV_DP_D, PORT_D);
> >   			if (I915_READ(CHV_HDMID) & SDVO_DETECTED || has_port)
> > -				intel_hdmi_init(dev_priv, CHV_HDMID, PORT_D);
> > +				intel_hdmi_init(i915, CHV_HDMID, PORT_D);
> >   		}
> > -		vlv_dsi_init(dev_priv);
> > -	} else if (IS_PINEVIEW(dev_priv)) {
> > -		intel_lvds_init(dev_priv);
> > -		intel_crt_init(dev_priv);
> > -	} else if (IS_GEN_RANGE(dev_priv, 3, 4)) {
> > +		vlv_dsi_init(i915);
> > +	} else if (IS_PINEVIEW(i915)) {
> > +		intel_lvds_init(i915);
> > +		intel_crt_init(i915);
> > +	} else if (IS_GEN_RANGE(i915, 3, 4)) {
> >   		bool found = false;
> > -		if (IS_MOBILE(dev_priv))
> > -			intel_lvds_init(dev_priv);
> > +		if (IS_MOBILE(i915))
> > +			intel_lvds_init(i915);
> > -		intel_crt_init(dev_priv);
> > +		intel_crt_init(i915);
> >   		if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) {
> >   			DRM_DEBUG_KMS("probing SDVOB\n");
> > -			found = intel_sdvo_init(dev_priv, GEN3_SDVOB, PORT_B);
> > -			if (!found && IS_G4X(dev_priv)) {
> > +			found = intel_sdvo_init(i915, GEN3_SDVOB, PORT_B);
> > +			if (!found && IS_G4X(i915)) {
> >   				DRM_DEBUG_KMS("probing HDMI on SDVOB\n");
> > -				intel_hdmi_init(dev_priv, GEN4_HDMIB, PORT_B);
> > +				intel_hdmi_init(i915, GEN4_HDMIB, PORT_B);
> >   			}
> > -			if (!found && IS_G4X(dev_priv))
> > -				intel_dp_init(dev_priv, DP_B, PORT_B);
> > +			if (!found && IS_G4X(i915))
> > +				intel_dp_init(i915, DP_B, PORT_B);
> >   		}
> >   		/* Before G4X SDVOC doesn't have its own detect register */
> >   		if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) {
> >   			DRM_DEBUG_KMS("probing SDVOC\n");
> > -			found = intel_sdvo_init(dev_priv, GEN3_SDVOC, PORT_C);
> > +			found = intel_sdvo_init(i915, GEN3_SDVOC, PORT_C);
> >   		}
> >   		if (!found && (I915_READ(GEN3_SDVOC) & SDVO_DETECTED)) {
> > -			if (IS_G4X(dev_priv)) {
> > +			if (IS_G4X(i915)) {
> >   				DRM_DEBUG_KMS("probing HDMI on SDVOC\n");
> > -				intel_hdmi_init(dev_priv, GEN4_HDMIC, PORT_C);
> > +				intel_hdmi_init(i915, GEN4_HDMIC, PORT_C);
> >   			}
> > -			if (IS_G4X(dev_priv))
> > -				intel_dp_init(dev_priv, DP_C, PORT_C);
> > +			if (IS_G4X(i915))
> > +				intel_dp_init(i915, DP_C, PORT_C);
> >   		}
> > -		if (IS_G4X(dev_priv) && (I915_READ(DP_D) & DP_DETECTED))
> > -			intel_dp_init(dev_priv, DP_D, PORT_D);
> > +		if (IS_G4X(i915) && (I915_READ(DP_D) & DP_DETECTED))
> > +			intel_dp_init(i915, DP_D, PORT_D);
> > -		if (SUPPORTS_TV(dev_priv))
> > -			intel_tv_init(dev_priv);
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > -		if (IS_I85X(dev_priv))
> > -			intel_lvds_init(dev_priv);
> > +		if (SUPPORTS_TV(i915))
> > +			intel_tv_init(i915);
> > +	} else if (IS_GEN(i915, 2)) {
> > +		if (IS_I85X(i915))
> > +			intel_lvds_init(i915);
> > -		intel_crt_init(dev_priv);
> > -		intel_dvo_init(dev_priv);
> > +		intel_crt_init(i915);
> > +		intel_dvo_init(i915);
> >   	}
> > -	intel_psr_init(dev_priv);
> > +	intel_psr_init(i915);
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		encoder->base.possible_crtcs = encoder->crtc_mask;
> >   		encoder->base.possible_clones =
> >   			intel_encoder_clones(encoder);
> >   	}
> > -	intel_init_pch_refclk(dev_priv);
> > +	intel_init_pch_refclk(i915);
> > -	drm_helper_move_panel_connectors_to_head(&dev_priv->drm);
> > +	drm_helper_move_panel_connectors_to_head(&i915->drm);
> >   }
> >   static void intel_user_framebuffer_destroy(struct drm_framebuffer *fb)
> > @@ -15379,7 +15379,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   				  struct drm_i915_gem_object *obj,
> >   				  struct drm_mode_fb_cmd2 *mode_cmd)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	struct drm_framebuffer *fb = &intel_fb->base;
> >   	u32 max_stride;
> >   	unsigned int tiling, stride;
> > @@ -15411,7 +15411,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   		}
> >   	}
> > -	if (!drm_any_plane_has_format(&dev_priv->drm,
> > +	if (!drm_any_plane_has_format(&i915->drm,
> >   				      mode_cmd->pixel_format,
> >   				      mode_cmd->modifier[0])) {
> >   		struct drm_format_name_buf format_name;
> > @@ -15427,13 +15427,13 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   	 * gen2/3 display engine uses the fence if present,
> >   	 * so the tiling mode must match the fb modifier exactly.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 4 &&
> > +	if (INTEL_GEN(i915) < 4 &&
> >   	    tiling != intel_fb_modifier_to_tiling(mode_cmd->modifier[0])) {
> >   		DRM_DEBUG_KMS("tiling_mode must match fb modifier exactly on gen2/3\n");
> >   		goto err;
> >   	}
> > -	max_stride = intel_fb_max_stride(dev_priv, mode_cmd->pixel_format,
> > +	max_stride = intel_fb_max_stride(i915, mode_cmd->pixel_format,
> >   					 mode_cmd->modifier[0]);
> >   	if (mode_cmd->pitches[0] > max_stride) {
> >   		DRM_DEBUG_KMS("%s pitch (%u) must be at most %d\n",
> > @@ -15457,7 +15457,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   	if (mode_cmd->offsets[0] != 0)
> >   		goto err;
> > -	drm_helper_mode_fill_fb_struct(&dev_priv->drm, fb, mode_cmd);
> > +	drm_helper_mode_fill_fb_struct(&i915->drm, fb, mode_cmd);
> >   	for (i = 0; i < fb->format->num_planes; i++) {
> >   		u32 stride_alignment;
> > @@ -15478,7 +15478,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   		 * require the entire fb to accommodate that to avoid
> >   		 * potential runtime errors at plane configuration time.
> >   		 */
> > -		if (IS_GEN(dev_priv, 9) && i == 0 && fb->width > 3840 &&
> > +		if (IS_GEN(i915, 9) && i == 0 && fb->width > 3840 &&
> >   		    is_ccs_modifier(fb->modifier))
> >   			stride_alignment *= 4;
> > @@ -15491,11 +15491,11 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb,
> >   		fb->obj[i] = &obj->base;
> >   	}
> > -	ret = intel_fill_fb_info(dev_priv, fb);
> > +	ret = intel_fill_fb_info(i915, fb);
> >   	if (ret)
> >   		goto err;
> > -	ret = drm_framebuffer_init(&dev_priv->drm, fb, &intel_fb_funcs);
> > +	ret = drm_framebuffer_init(&i915->drm, fb, &intel_fb_funcs);
> >   	if (ret) {
> >   		DRM_ERROR("framebuffer init failed %d\n", ret);
> >   		goto err;
> > @@ -15545,7 +15545,7 @@ static enum drm_mode_status
> >   intel_mode_valid(struct drm_device *dev,
> >   		 const struct drm_display_mode *mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	int hdisplay_max, htotal_max;
> >   	int vdisplay_max, vtotal_max;
> > @@ -15578,13 +15578,13 @@ intel_mode_valid(struct drm_device *dev,
> >   			   DRM_MODE_FLAG_CLKDIV2))
> >   		return MODE_BAD;
> > -	if (INTEL_GEN(dev_priv) >= 9 ||
> > -	    IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 9 ||
> > +	    IS_BROADWELL(i915) || IS_HASWELL(i915)) {
> >   		hdisplay_max = 8192; /* FDI max 4096 handled elsewhere */
> >   		vdisplay_max = 4096;
> >   		htotal_max = 8192;
> >   		vtotal_max = 8192;
> > -	} else if (INTEL_GEN(dev_priv) >= 3) {
> > +	} else if (INTEL_GEN(i915) >= 3) {
> >   		hdisplay_max = 4096;
> >   		vdisplay_max = 4096;
> >   		htotal_max = 8192;
> > @@ -15625,113 +15625,113 @@ static const struct drm_mode_config_funcs intel_mode_funcs = {
> >   /**
> >    * intel_init_display_hooks - initialize the display modesetting hooks
> > - * @dev_priv: device private
> > + * @i915: device private
> >    */
> > -void intel_init_display_hooks(struct drm_i915_private *dev_priv)
> > +void intel_init_display_hooks(struct drm_i915_private *i915)
> >   {
> > -	intel_init_cdclk_hooks(dev_priv);
> > +	intel_init_cdclk_hooks(i915);
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > -		dev_priv->display.get_pipe_config = haswell_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +	if (INTEL_GEN(i915) >= 9) {
> > +		i915->display.get_pipe_config = haswell_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			skylake_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock =
> > +		i915->display.crtc_compute_clock =
> >   			haswell_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = haswell_crtc_enable;
> > -		dev_priv->display.crtc_disable = haswell_crtc_disable;
> > -	} else if (HAS_DDI(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = haswell_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_enable = haswell_crtc_enable;
> > +		i915->display.crtc_disable = haswell_crtc_disable;
> > +	} else if (HAS_DDI(i915)) {
> > +		i915->display.get_pipe_config = haswell_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock =
> > +		i915->display.crtc_compute_clock =
> >   			haswell_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = haswell_crtc_enable;
> > -		dev_priv->display.crtc_disable = haswell_crtc_disable;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = ironlake_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_enable = haswell_crtc_enable;
> > +		i915->display.crtc_disable = haswell_crtc_disable;
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> > +		i915->display.get_pipe_config = ironlake_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock =
> > +		i915->display.crtc_compute_clock =
> >   			ironlake_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = ironlake_crtc_enable;
> > -		dev_priv->display.crtc_disable = ironlake_crtc_disable;
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_enable = ironlake_crtc_enable;
> > +		i915->display.crtc_disable = ironlake_crtc_disable;
> > +	} else if (IS_CHERRYVIEW(i915)) {
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = chv_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = valleyview_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_compute_clock = chv_crtc_compute_clock;
> > +		i915->display.crtc_enable = valleyview_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> > +	} else if (IS_VALLEYVIEW(i915)) {
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = vlv_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = valleyview_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > -	} else if (IS_G4X(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_compute_clock = vlv_crtc_compute_clock;
> > +		i915->display.crtc_enable = valleyview_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> > +	} else if (IS_G4X(i915)) {
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = g4x_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = i9xx_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > -	} else if (IS_PINEVIEW(dev_priv)) {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_compute_clock = g4x_crtc_compute_clock;
> > +		i915->display.crtc_enable = i9xx_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> > +	} else if (IS_PINEVIEW(i915)) {
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = pnv_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = i9xx_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > -	} else if (!IS_GEN(dev_priv, 2)) {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.crtc_compute_clock = pnv_crtc_compute_clock;
> > +		i915->display.crtc_enable = i9xx_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> > +	} else if (!IS_GEN(i915, 2)) {
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = i9xx_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = i9xx_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > +		i915->display.crtc_compute_clock = i9xx_crtc_compute_clock;
> > +		i915->display.crtc_enable = i9xx_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> >   	} else {
> > -		dev_priv->display.get_pipe_config = i9xx_get_pipe_config;
> > -		dev_priv->display.get_initial_plane_config =
> > +		i915->display.get_pipe_config = i9xx_get_pipe_config;
> > +		i915->display.get_initial_plane_config =
> >   			i9xx_get_initial_plane_config;
> > -		dev_priv->display.crtc_compute_clock = i8xx_crtc_compute_clock;
> > -		dev_priv->display.crtc_enable = i9xx_crtc_enable;
> > -		dev_priv->display.crtc_disable = i9xx_crtc_disable;
> > +		i915->display.crtc_compute_clock = i8xx_crtc_compute_clock;
> > +		i915->display.crtc_enable = i9xx_crtc_enable;
> > +		i915->display.crtc_disable = i9xx_crtc_disable;
> >   	}
> > -	if (IS_GEN(dev_priv, 5)) {
> > -		dev_priv->display.fdi_link_train = ironlake_fdi_link_train;
> > -	} else if (IS_GEN(dev_priv, 6)) {
> > -		dev_priv->display.fdi_link_train = gen6_fdi_link_train;
> > -	} else if (IS_IVYBRIDGE(dev_priv)) {
> > +	if (IS_GEN(i915, 5)) {
> > +		i915->display.fdi_link_train = ironlake_fdi_link_train;
> > +	} else if (IS_GEN(i915, 6)) {
> > +		i915->display.fdi_link_train = gen6_fdi_link_train;
> > +	} else if (IS_IVYBRIDGE(i915)) {
> >   		/* FIXME: detect B0+ stepping and use auto training */
> > -		dev_priv->display.fdi_link_train = ivb_manual_fdi_link_train;
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > -		dev_priv->display.fdi_link_train = hsw_fdi_link_train;
> > +		i915->display.fdi_link_train = ivb_manual_fdi_link_train;
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> > +		i915->display.fdi_link_train = hsw_fdi_link_train;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		dev_priv->display.update_crtcs = skl_update_crtcs;
> > +	if (INTEL_GEN(i915) >= 9)
> > +		i915->display.update_crtcs = skl_update_crtcs;
> >   	else
> > -		dev_priv->display.update_crtcs = intel_update_crtcs;
> > +		i915->display.update_crtcs = intel_update_crtcs;
> >   }
> > -static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *dev_priv)
> > +static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *i915)
> >   {
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		return VLV_VGACNTRL;
> > -	else if (INTEL_GEN(dev_priv) >= 5)
> > +	else if (INTEL_GEN(i915) >= 5)
> >   		return CPU_VGACNTRL;
> >   	else
> >   		return VGACNTRL;
> >   }
> >   /* Disable the VGA plane that we never use */
> > -static void i915_disable_vga(struct drm_i915_private *dev_priv)
> > +static void i915_disable_vga(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u8 sr1;
> > -	i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv);
> > +	i915_reg_t vga_reg = i915_vgacntrl_reg(i915);
> >   	/* WaEnableVGAAccessThroughIOPort:ctg,elk,ilk,snb,ivb,vlv,hsw */
> >   	vga_get_uninterruptible(pdev, VGA_RSRC_LEGACY_IO);
> > @@ -15747,11 +15747,11 @@ static void i915_disable_vga(struct drm_i915_private *dev_priv)
> >   void intel_modeset_init_hw(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > -	dev_priv->cdclk.logical = dev_priv->cdclk.actual = dev_priv->cdclk.hw;
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> > +	i915->cdclk.logical = i915->cdclk.actual = i915->cdclk.hw;
> >   }
> >   /*
> > @@ -15766,7 +15766,7 @@ void intel_modeset_init_hw(struct drm_device *dev)
> >    */
> >   static void sanitize_watermarks(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_atomic_state *state;
> >   	struct intel_atomic_state *intel_state;
> >   	struct drm_crtc *crtc;
> > @@ -15776,7 +15776,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> >   	int i;
> >   	/* Only supported on platforms that use atomic watermark design */
> > -	if (!dev_priv->display.optimize_watermarks)
> > +	if (!i915->display.optimize_watermarks)
> >   		return;
> >   	/*
> > @@ -15804,7 +15804,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> >   	 * intermediate watermarks (since we don't trust the current
> >   	 * watermarks).
> >   	 */
> > -	if (!HAS_GMCH(dev_priv))
> > +	if (!HAS_GMCH(i915))
> >   		intel_state->skip_intermediate_wm = true;
> >   	ret = intel_atomic_check(dev, state);
> > @@ -15829,7 +15829,7 @@ static void sanitize_watermarks(struct drm_device *dev)
> >   		struct intel_crtc_state *cs = to_intel_crtc_state(cstate);
> >   		cs->wm.need_postvbl_update = true;
> > -		dev_priv->display.optimize_watermarks(intel_state, cs);
> > +		i915->display.optimize_watermarks(intel_state, cs);
> >   		to_intel_crtc_state(crtc->state)->wm = cs->wm;
> >   	}
> > @@ -15841,20 +15841,20 @@ static void sanitize_watermarks(struct drm_device *dev)
> >   	drm_modeset_acquire_fini(&ctx);
> >   }
> > -static void intel_update_fdi_pll_freq(struct drm_i915_private *dev_priv)
> > +static void intel_update_fdi_pll_freq(struct drm_i915_private *i915)
> >   {
> > -	if (IS_GEN(dev_priv, 5)) {
> > +	if (IS_GEN(i915, 5)) {
> >   		u32 fdi_pll_clk =
> >   			I915_READ(FDI_PLL_BIOS_0) & FDI_PLL_FB_CLOCK_MASK;
> > -		dev_priv->fdi_pll_freq = (fdi_pll_clk + 2) * 10000;
> > -	} else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) {
> > -		dev_priv->fdi_pll_freq = 270000;
> > +		i915->fdi_pll_freq = (fdi_pll_clk + 2) * 10000;
> > +	} else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) {
> > +		i915->fdi_pll_freq = 270000;
> >   	} else {
> >   		return;
> >   	}
> > -	DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", dev_priv->fdi_pll_freq);
> > +	DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", i915->fdi_pll_freq);
> >   }
> >   static int intel_initial_commit(struct drm_device *dev)
> > @@ -15915,17 +15915,17 @@ static int intel_initial_commit(struct drm_device *dev)
> >   int intel_modeset_init(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	enum pipe pipe;
> >   	struct intel_crtc *crtc;
> >   	int ret;
> > -	dev_priv->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0);
> > +	i915->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0);
> >   	drm_mode_config_init(dev);
> > -	ret = intel_bw_init(dev_priv);
> > +	ret = intel_bw_init(i915);
> >   	if (ret)
> >   		return ret;
> > @@ -15939,15 +15939,15 @@ int intel_modeset_init(struct drm_device *dev)
> >   	dev->mode_config.funcs = &intel_mode_funcs;
> > -	init_llist_head(&dev_priv->atomic_helper.free_list);
> > -	INIT_WORK(&dev_priv->atomic_helper.free_work,
> > +	init_llist_head(&i915->atomic_helper.free_list);
> > +	INIT_WORK(&i915->atomic_helper.free_work,
> >   		  intel_atomic_helper_free_state_worker);
> > -	intel_init_quirks(dev_priv);
> > +	intel_init_quirks(i915);
> > -	intel_fbc_init(dev_priv);
> > +	intel_fbc_init(i915);
> > -	intel_init_pm(dev_priv);
> > +	intel_init_pm(i915);
> >   	/*
> >   	 * There may be no VBT; and if the BIOS enabled SSC we can
> > @@ -15955,15 +15955,15 @@ int intel_modeset_init(struct drm_device *dev)
> >   	 * BIOS isn't using it, don't assume it will work even if the VBT
> >   	 * indicates as much.
> >   	 */
> > -	if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) {
> >   		bool bios_lvds_use_ssc = !!(I915_READ(PCH_DREF_CONTROL) &
> >   					    DREF_SSC1_ENABLE);
> > -		if (dev_priv->vbt.lvds_use_ssc != bios_lvds_use_ssc) {
> > +		if (i915->vbt.lvds_use_ssc != bios_lvds_use_ssc) {
> >   			DRM_DEBUG_KMS("SSC %sabled by BIOS, overriding VBT which says %sabled\n",
> >   				     bios_lvds_use_ssc ? "en" : "dis",
> > -				     dev_priv->vbt.lvds_use_ssc ? "en" : "dis");
> > -			dev_priv->vbt.lvds_use_ssc = bios_lvds_use_ssc;
> > +				     i915->vbt.lvds_use_ssc ? "en" : "dis");
> > +			i915->vbt.lvds_use_ssc = bios_lvds_use_ssc;
> >   		}
> >   	}
> > @@ -15971,13 +15971,13 @@ int intel_modeset_init(struct drm_device *dev)
> >   	 * Maximum framebuffer dimensions, chosen to match
> >   	 * the maximum render engine surface size on gen4+.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 7) {
> > +	if (INTEL_GEN(i915) >= 7) {
> >   		dev->mode_config.max_width = 16384;
> >   		dev->mode_config.max_height = 16384;
> > -	} else if (INTEL_GEN(dev_priv) >= 4) {
> > +	} else if (INTEL_GEN(i915) >= 4) {
> >   		dev->mode_config.max_width = 8192;
> >   		dev->mode_config.max_height = 8192;
> > -	} else if (IS_GEN(dev_priv, 3)) {
> > +	} else if (IS_GEN(i915, 3)) {
> >   		dev->mode_config.max_width = 4096;
> >   		dev->mode_config.max_height = 4096;
> >   	} else {
> > @@ -15985,10 +15985,10 @@ int intel_modeset_init(struct drm_device *dev)
> >   		dev->mode_config.max_height = 2048;
> >   	}
> > -	if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) {
> > -		dev->mode_config.cursor_width = IS_I845G(dev_priv) ? 64 : 512;
> > +	if (IS_I845G(i915) || IS_I865G(i915)) {
> > +		dev->mode_config.cursor_width = IS_I845G(i915) ? 64 : 512;
> >   		dev->mode_config.cursor_height = 1023;
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > +	} else if (IS_GEN(i915, 2)) {
> >   		dev->mode_config.cursor_width = 64;
> >   		dev->mode_config.cursor_height = 64;
> >   	} else {
> > @@ -15999,11 +15999,11 @@ int intel_modeset_init(struct drm_device *dev)
> >   	dev->mode_config.fb_base = ggtt->gmadr.start;
> >   	DRM_DEBUG_KMS("%d display pipe%s available.\n",
> > -		      INTEL_INFO(dev_priv)->num_pipes,
> > -		      INTEL_INFO(dev_priv)->num_pipes > 1 ? "s" : "");
> > +		      INTEL_INFO(i915)->num_pipes,
> > +		      INTEL_INFO(i915)->num_pipes > 1 ? "s" : "");
> > -	for_each_pipe(dev_priv, pipe) {
> > -		ret = intel_crtc_init(dev_priv, pipe);
> > +	for_each_pipe(i915, pipe) {
> > +		ret = intel_crtc_init(i915, pipe);
> >   		if (ret) {
> >   			drm_mode_config_cleanup(dev);
> >   			return ret;
> > @@ -16011,19 +16011,19 @@ int intel_modeset_init(struct drm_device *dev)
> >   	}
> >   	intel_shared_dpll_init(dev);
> > -	intel_update_fdi_pll_freq(dev_priv);
> > +	intel_update_fdi_pll_freq(i915);
> > -	intel_update_czclk(dev_priv);
> > +	intel_update_czclk(i915);
> >   	intel_modeset_init_hw(dev);
> > -	intel_hdcp_component_init(dev_priv);
> > +	intel_hdcp_component_init(i915);
> > -	if (dev_priv->max_cdclk_freq == 0)
> > -		intel_update_max_cdclk(dev_priv);
> > +	if (i915->max_cdclk_freq == 0)
> > +		intel_update_max_cdclk(i915);
> >   	/* Just disable it once at startup */
> > -	i915_disable_vga(dev_priv);
> > -	intel_setup_outputs(dev_priv);
> > +	i915_disable_vga(i915);
> > +	intel_setup_outputs(i915);
> >   	drm_modeset_lock_all(dev);
> >   	intel_modeset_setup_hw_state(dev, dev->mode_config.acquire_ctx);
> > @@ -16042,7 +16042,7 @@ int intel_modeset_init(struct drm_device *dev)
> >   		 * can even allow for smooth boot transitions if the BIOS
> >   		 * fb is large enough for the active pipe configuration.
> >   		 */
> > -		dev_priv->display.get_initial_plane_config(crtc,
> > +		i915->display.get_initial_plane_config(crtc,
> >   							   &plane_config);
> >   		/*
> > @@ -16057,7 +16057,7 @@ int intel_modeset_init(struct drm_device *dev)
> >   	 * Note that we need to do this after reconstructing the BIOS fb's
> >   	 * since the watermark calculation done here will use pstate->fb.
> >   	 */
> > -	if (!HAS_GMCH(dev_priv))
> > +	if (!HAS_GMCH(i915))
> >   		sanitize_watermarks(dev);
> >   	/*
> > @@ -16073,9 +16073,9 @@ int intel_modeset_init(struct drm_device *dev)
> >   	return 0;
> >   }
> > -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	/* 640x480 at 60Hz, ~25175 kHz */
> >   	struct dpll clock = {
> >   		.m1 = 18,
> > @@ -16143,9 +16143,9 @@ void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   	intel_wait_for_pipe_scanline_moving(crtc);
> >   }
> > -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	DRM_DEBUG_KMS("disabling pipe %c due to force quirk\n",
> >   		      pipe_name(pipe));
> > @@ -16166,14 +16166,14 @@ void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   }
> >   static void
> > -intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv)
> > +intel_sanitize_plane_mapping(struct drm_i915_private *i915)
> >   {
> >   	struct intel_crtc *crtc;
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		return;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_plane *plane =
> >   			to_intel_plane(crtc->base.primary);
> >   		struct intel_crtc *plane_crtc;
> > @@ -16188,7 +16188,7 @@ intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv)
> >   		DRM_DEBUG_KMS("[PLANE:%d:%s] attached to the wrong pipe, disabling plane\n",
> >   			      plane->base.base.id, plane->base.name);
> > -		plane_crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +		plane_crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   		intel_plane_disable_noatomic(plane_crtc, plane);
> >   	}
> >   }
> > @@ -16215,18 +16215,18 @@ static struct intel_connector *intel_encoder_find_connector(struct intel_encoder
> >   	return NULL;
> >   }
> > -static bool has_pch_trancoder(struct drm_i915_private *dev_priv,
> > +static bool has_pch_trancoder(struct drm_i915_private *i915,
> >   			      enum pipe pch_transcoder)
> >   {
> > -	return HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) ||
> > -		(HAS_PCH_LPT_H(dev_priv) && pch_transcoder == PIPE_A);
> > +	return HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) ||
> > +		(HAS_PCH_LPT_H(i915) && pch_transcoder == PIPE_A);
> >   }
> >   static void intel_sanitize_crtc(struct intel_crtc *crtc,
> >   				struct drm_modeset_acquire_ctx *ctx)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc_state *crtc_state = to_intel_crtc_state(crtc->base.state);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> > @@ -16255,7 +16255,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> >   		 * Disable any background color set by the BIOS, but enable the
> >   		 * gamma and CSC to match how we program our planes.
> >   		 */
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			I915_WRITE(SKL_BOTTOM_COLOR(crtc->pipe),
> >   				   SKL_BOTTOM_COLOR_GAMMA_ENABLE |
> >   				   SKL_BOTTOM_COLOR_CSC_ENABLE);
> > @@ -16266,7 +16266,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> >   	if (crtc_state->base.active && !intel_crtc_has_encoders(crtc))
> >   		intel_crtc_disable_noatomic(&crtc->base, ctx);
> > -	if (crtc_state->base.active || HAS_GMCH(dev_priv)) {
> > +	if (crtc_state->base.active || HAS_GMCH(i915)) {
> >   		/*
> >   		 * We start out with underrun reporting disabled to avoid races.
> >   		 * For correct bookkeeping mark this on active crtcs.
> > @@ -16290,14 +16290,14 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc,
> >   		 * PCH transcoders B and C would prevent enabling the south
> >   		 * error interrupt (see cpt_can_enable_serr_int()).
> >   		 */
> > -		if (has_pch_trancoder(dev_priv, crtc->pipe))
> > +		if (has_pch_trancoder(i915, crtc->pipe))
> >   			crtc->pch_fifo_underrun_disabled = true;
> >   	}
> >   }
> >   static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	/*
> >   	 * Some SNB BIOSen (eg. ASUS K53SV) are known to misprogram
> > @@ -16309,7 +16309,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
> >   	 * without several WARNs, but for now let's take the easy
> >   	 * road.
> >   	 */
> > -	return IS_GEN(dev_priv, 6) &&
> > +	return IS_GEN(i915, 6) &&
> >   		crtc_state->base.active &&
> >   		crtc_state->shared_dpll &&
> >   		crtc_state->port_clock == 0;
> > @@ -16317,7 +16317,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state)
> >   static void intel_sanitize_encoder(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_connector *connector;
> >   	struct intel_crtc *crtc = to_intel_crtc(encoder->base.crtc);
> >   	struct intel_crtc_state *crtc_state = crtc ?
> > @@ -16378,21 +16378,21 @@ static void intel_sanitize_encoder(struct intel_encoder *encoder)
> >   	/* notify opregion of the sanitized encoder state */
> >   	intel_opregion_notify_encoder(encoder, connector && has_active_crtc);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		icl_sanitize_encoder_pll_mapping(encoder);
> >   }
> > -void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv)
> > +void i915_redisable_vga_power_on(struct drm_i915_private *i915)
> >   {
> > -	i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv);
> > +	i915_reg_t vga_reg = i915_vgacntrl_reg(i915);
> >   	if (!(I915_READ(vga_reg) & VGA_DISP_DISABLE)) {
> >   		DRM_DEBUG_KMS("Something enabled VGA plane, disabling it\n");
> > -		i915_disable_vga(dev_priv);
> > +		i915_disable_vga(i915);
> >   	}
> >   }
> > -void i915_redisable_vga(struct drm_i915_private *dev_priv)
> > +void i915_redisable_vga(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref;
> > @@ -16405,23 +16405,23 @@ void i915_redisable_vga(struct drm_i915_private *dev_priv)
> >   	 * follow the "don't touch the power well if we don't need it" policy
> >   	 * the rest of the driver uses.
> >   	 */
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_VGA);
> >   	if (!wakeref)
> >   		return;
> > -	i915_redisable_vga_power_on(dev_priv);
> > +	i915_redisable_vga_power_on(i915);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_VGA, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_VGA, wakeref);
> >   }
> >   /* FIXME read out full plane state for all planes */
> > -static void readout_plane_state(struct drm_i915_private *dev_priv)
> > +static void readout_plane_state(struct drm_i915_private *i915)
> >   {
> >   	struct intel_plane *plane;
> >   	struct intel_crtc *crtc;
> > -	for_each_intel_plane(&dev_priv->drm, plane) {
> > +	for_each_intel_plane(&i915->drm, plane) {
> >   		struct intel_plane_state *plane_state =
> >   			to_intel_plane_state(plane->base.state);
> >   		struct intel_crtc_state *crtc_state;
> > @@ -16430,7 +16430,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
> >   		visible = plane->get_hw_state(plane, &pipe);
> > -		crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +		crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   		crtc_state = to_intel_crtc_state(crtc->base.state);
> >   		intel_set_plane_visible(crtc_state, plane_state, visible);
> > @@ -16440,7 +16440,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
> >   			      enableddisabled(visible), pipe_name(pipe));
> >   	}
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> > @@ -16450,7 +16450,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv)
> >   static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum pipe pipe;
> >   	struct intel_crtc *crtc;
> >   	struct intel_encoder *encoder;
> > @@ -16458,7 +16458,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   	struct drm_connector_list_iter conn_iter;
> >   	int i;
> > -	dev_priv->active_crtcs = 0;
> > +	i915->active_crtcs = 0;
> >   	for_each_intel_crtc(dev, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> > @@ -16469,25 +16469,25 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   		__drm_atomic_helper_crtc_reset(&crtc->base, &crtc_state->base);
> >   		crtc_state->base.active = crtc_state->base.enable =
> > -			dev_priv->display.get_pipe_config(crtc, crtc_state);
> > +			i915->display.get_pipe_config(crtc, crtc_state);
> >   		crtc->base.enabled = crtc_state->base.enable;
> >   		crtc->active = crtc_state->base.active;
> >   		if (crtc_state->base.active)
> > -			dev_priv->active_crtcs |= 1 << crtc->pipe;
> > +			i915->active_crtcs |= 1 << crtc->pipe;
> >   		DRM_DEBUG_KMS("[CRTC:%d:%s] hw state readout: %s\n",
> >   			      crtc->base.base.id, crtc->base.name,
> >   			      enableddisabled(crtc_state->base.active));
> >   	}
> > -	readout_plane_state(dev_priv);
> > +	readout_plane_state(i915);
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > -		struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> > +		struct intel_shared_dpll *pll = &i915->shared_dplls[i];
> > -		pll->on = pll->info->funcs->get_hw_state(dev_priv, pll,
> > +		pll->on = pll->info->funcs->get_hw_state(i915, pll,
> >   							&pll->state.hw_state);
> >   		pll->state.crtc_mask = 0;
> >   		for_each_intel_crtc(dev, crtc) {
> > @@ -16510,7 +16510,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   		if (encoder->get_hw_state(encoder, &pipe)) {
> >   			struct intel_crtc_state *crtc_state;
> > -			crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +			crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   			crtc_state = to_intel_crtc_state(crtc->base.state);
> >   			encoder->base.crtc = &crtc->base;
> > @@ -16558,7 +16558,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   	for_each_intel_crtc(dev, crtc) {
> >   		struct intel_bw_state *bw_state =
> > -			to_intel_bw_state(dev_priv->bw_obj.state);
> > +			to_intel_bw_state(i915->bw_obj.state);
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		struct intel_plane *plane;
> > @@ -16585,7 +16585,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   			intel_crtc_compute_pixel_rate(crtc_state);
> > -			if (dev_priv->display.modeset_calc_cdclk) {
> > +			if (i915->display.modeset_calc_cdclk) {
> >   				min_cdclk = intel_crtc_compute_min_cdclk(crtc_state);
> >   				if (WARN_ON(min_cdclk < 0))
> >   					min_cdclk = 0;
> > @@ -16596,11 +16596,11 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   			update_scanline_offset(crtc_state);
> >   		}
> > -		dev_priv->min_cdclk[crtc->pipe] = min_cdclk;
> > -		dev_priv->min_voltage_level[crtc->pipe] =
> > +		i915->min_cdclk[crtc->pipe] = min_cdclk;
> > +		i915->min_voltage_level[crtc->pipe] =
> >   			crtc_state->min_voltage_level;
> > -		for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +		for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   			const struct intel_plane_state *plane_state =
> >   				to_intel_plane_state(plane->base.state);
> > @@ -16615,16 +16615,16 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev)
> >   		intel_bw_crtc_update(bw_state, crtc_state);
> > -		intel_pipe_config_sanity_check(dev_priv, crtc_state);
> > +		intel_pipe_config_sanity_check(i915, crtc_state);
> >   	}
> >   }
> >   static void
> > -get_encoder_power_domains(struct drm_i915_private *dev_priv)
> > +get_encoder_power_domains(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		struct intel_crtc_state *crtc_state;
> >   		if (!encoder->get_power_domains)
> > @@ -16642,14 +16642,14 @@ get_encoder_power_domains(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void intel_early_display_was(struct drm_i915_private *dev_priv)
> > +static void intel_early_display_was(struct drm_i915_private *i915)
> >   {
> >   	/* Display WA #1185 WaDisableDARBFClkGating:cnl,glk */
> > -	if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv))
> > +	if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915))
> >   		I915_WRITE(GEN9_CLKGATE_DIS_0, I915_READ(GEN9_CLKGATE_DIS_0) |
> >   			   DARBF_GATING_DIS);
> > -	if (IS_HASWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915)) {
> >   		/*
> >   		 * WaRsPkgCStateDisplayPMReq:hsw
> >   		 * System hang if this isn't done before disabling all planes!
> > @@ -16659,7 +16659,7 @@ static void intel_early_display_was(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv,
> > +static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *i915,
> >   				       enum port port, i915_reg_t hdmi_reg)
> >   {
> >   	u32 val = I915_READ(hdmi_reg);
> > @@ -16677,7 +16677,7 @@ static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(hdmi_reg, val);
> >   }
> > -static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv,
> > +static void ibx_sanitize_pch_dp_port(struct drm_i915_private *i915,
> >   				     enum port port, i915_reg_t dp_reg)
> >   {
> >   	u32 val = I915_READ(dp_reg);
> > @@ -16695,7 +16695,7 @@ static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(dp_reg, val);
> >   }
> > -static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv)
> > +static void ibx_sanitize_pch_ports(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * The BIOS may select transcoder B on some of the PCH
> > @@ -16708,14 +16708,14 @@ static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv)
> >   	 * (see. intel_dp_link_down(), intel_disable_hdmi(),
> >   	 * intel_disable_sdvo()).
> >   	 */
> > -	ibx_sanitize_pch_dp_port(dev_priv, PORT_B, PCH_DP_B);
> > -	ibx_sanitize_pch_dp_port(dev_priv, PORT_C, PCH_DP_C);
> > -	ibx_sanitize_pch_dp_port(dev_priv, PORT_D, PCH_DP_D);
> > +	ibx_sanitize_pch_dp_port(i915, PORT_B, PCH_DP_B);
> > +	ibx_sanitize_pch_dp_port(i915, PORT_C, PCH_DP_C);
> > +	ibx_sanitize_pch_dp_port(i915, PORT_D, PCH_DP_D);
> >   	/* PCH SDVOB multiplex with HDMIB */
> > -	ibx_sanitize_pch_hdmi_port(dev_priv, PORT_B, PCH_HDMIB);
> > -	ibx_sanitize_pch_hdmi_port(dev_priv, PORT_C, PCH_HDMIC);
> > -	ibx_sanitize_pch_hdmi_port(dev_priv, PORT_D, PCH_HDMID);
> > +	ibx_sanitize_pch_hdmi_port(i915, PORT_B, PCH_HDMIB);
> > +	ibx_sanitize_pch_hdmi_port(i915, PORT_C, PCH_HDMIC);
> > +	ibx_sanitize_pch_hdmi_port(i915, PORT_D, PCH_HDMID);
> >   }
> >   /* Scan out the current hw modeset state,
> > @@ -16725,29 +16725,29 @@ static void
> >   intel_modeset_setup_hw_state(struct drm_device *dev,
> >   			     struct drm_modeset_acquire_ctx *ctx)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc_state *crtc_state;
> >   	struct intel_encoder *encoder;
> >   	struct intel_crtc *crtc;
> >   	intel_wakeref_t wakeref;
> >   	int i;
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT);
> > -	intel_early_display_was(dev_priv);
> > +	intel_early_display_was(i915);
> >   	intel_modeset_readout_hw_state(dev);
> >   	/* HW state is read out, now we need to sanitize this mess. */
> > -	get_encoder_power_domains(dev_priv);
> > +	get_encoder_power_domains(i915);
> > -	if (HAS_PCH_IBX(dev_priv))
> > -		ibx_sanitize_pch_ports(dev_priv);
> > +	if (HAS_PCH_IBX(i915))
> > +		ibx_sanitize_pch_ports(i915);
> >   	/*
> >   	 * intel_sanitize_plane_mapping() may need to do vblank
> >   	 * waits, so we need vblank interrupts restored beforehand.
> >   	 */
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		crtc_state = to_intel_crtc_state(crtc->base.state);
> >   		drm_crtc_vblank_reset(&crtc->base);
> > @@ -16756,12 +16756,12 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> >   			intel_crtc_vblank_on(crtc_state);
> >   	}
> > -	intel_sanitize_plane_mapping(dev_priv);
> > +	intel_sanitize_plane_mapping(i915);
> >   	for_each_intel_encoder(dev, encoder)
> >   		intel_sanitize_encoder(encoder);
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		crtc_state = to_intel_crtc_state(crtc->base.state);
> >   		intel_sanitize_crtc(crtc, ctx);
> >   		intel_dump_pipe_config(crtc_state, NULL, "[setup_hw_state]");
> > @@ -16769,8 +16769,8 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> >   	intel_modeset_update_connector_atomic_state(dev);
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > -		struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> > +		struct intel_shared_dpll *pll = &i915->shared_dplls[i];
> >   		if (!pll->on || pll->active_mask)
> >   			continue;
> > @@ -16778,20 +16778,20 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> >   		DRM_DEBUG_KMS("%s enabled but not in use, disabling\n",
> >   			      pll->info->name);
> > -		pll->info->funcs->disable(dev_priv, pll);
> > +		pll->info->funcs->disable(i915, pll);
> >   		pll->on = false;
> >   	}
> > -	if (IS_G4X(dev_priv)) {
> > -		g4x_wm_get_hw_state(dev_priv);
> > -		g4x_wm_sanitize(dev_priv);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		vlv_wm_get_hw_state(dev_priv);
> > -		vlv_wm_sanitize(dev_priv);
> > -	} else if (INTEL_GEN(dev_priv) >= 9) {
> > -		skl_wm_get_hw_state(dev_priv);
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > -		ilk_wm_get_hw_state(dev_priv);
> > +	if (IS_G4X(i915)) {
> > +		g4x_wm_get_hw_state(i915);
> > +		g4x_wm_sanitize(i915);
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		vlv_wm_get_hw_state(i915);
> > +		vlv_wm_sanitize(i915);
> > +	} else if (INTEL_GEN(i915) >= 9) {
> > +		skl_wm_get_hw_state(i915);
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> > +		ilk_wm_get_hw_state(i915);
> >   	}
> >   	for_each_intel_crtc(dev, crtc) {
> > @@ -16800,22 +16800,22 @@ intel_modeset_setup_hw_state(struct drm_device *dev,
> >   		crtc_state = to_intel_crtc_state(crtc->base.state);
> >   		put_domains = modeset_get_crtc_power_domains(&crtc->base, crtc_state);
> >   		if (WARN_ON(put_domains))
> > -			modeset_put_power_domains(dev_priv, put_domains);
> > +			modeset_put_power_domains(i915, put_domains);
> >   	}
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
> > -	intel_fbc_init_pipe_state(dev_priv);
> > +	intel_fbc_init_pipe_state(i915);
> >   }
> >   void intel_display_resume(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct drm_atomic_state *state = dev_priv->modeset_restore_state;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct drm_atomic_state *state = i915->modeset_restore_state;
> >   	struct drm_modeset_acquire_ctx ctx;
> >   	int ret;
> > -	dev_priv->modeset_restore_state = NULL;
> > +	i915->modeset_restore_state = NULL;
> >   	if (state)
> >   		state->acquire_ctx = &ctx;
> > @@ -16832,7 +16832,7 @@ void intel_display_resume(struct drm_device *dev)
> >   	if (!ret)
> >   		ret = __intel_display_resume(dev, state, &ctx);
> > -	intel_enable_ipc(dev_priv);
> > +	intel_enable_ipc(i915);
> >   	drm_modeset_drop_locks(&ctx);
> >   	drm_modeset_acquire_fini(&ctx);
> > @@ -16862,19 +16862,19 @@ static void intel_hpd_poll_fini(struct drm_device *dev)
> >   void intel_modeset_cleanup(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	flush_workqueue(dev_priv->modeset_wq);
> > +	flush_workqueue(i915->modeset_wq);
> > -	flush_work(&dev_priv->atomic_helper.free_work);
> > -	WARN_ON(!llist_empty(&dev_priv->atomic_helper.free_list));
> > +	flush_work(&i915->atomic_helper.free_work);
> > +	WARN_ON(!llist_empty(&i915->atomic_helper.free_list));
> >   	/*
> >   	 * Interrupts and polling as the first thing to avoid creating havoc.
> >   	 * Too much stuff here (turning of connectors, ...) would
> >   	 * experience fancy races otherwise.
> >   	 */
> > -	intel_irq_uninstall(dev_priv);
> > +	intel_irq_uninstall(i915);
> >   	/*
> >   	 * Due to the hpd irq storm handling the hotplug work can re-arm the
> > @@ -16883,37 +16883,37 @@ void intel_modeset_cleanup(struct drm_device *dev)
> >   	intel_hpd_poll_fini(dev);
> >   	/* poll work can call into fbdev, hence clean that up afterwards */
> > -	intel_fbdev_fini(dev_priv);
> > +	intel_fbdev_fini(i915);
> >   	intel_unregister_dsm_handler();
> > -	intel_fbc_global_disable(dev_priv);
> > +	intel_fbc_global_disable(i915);
> >   	/* flush any delayed tasks or pending work */
> >   	flush_scheduled_work();
> > -	intel_hdcp_component_fini(dev_priv);
> > +	intel_hdcp_component_fini(i915);
> >   	drm_mode_config_cleanup(dev);
> > -	intel_overlay_cleanup(dev_priv);
> > +	intel_overlay_cleanup(i915);
> > -	intel_gmbus_teardown(dev_priv);
> > +	intel_gmbus_teardown(i915);
> > -	destroy_workqueue(dev_priv->modeset_wq);
> > +	destroy_workqueue(i915->modeset_wq);
> > -	intel_fbc_cleanup_cfb(dev_priv);
> > +	intel_fbc_cleanup_cfb(i915);
> >   }
> >   /*
> >    * set vga decode state - true == enable VGA decode
> >    */
> > -int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state)
> > +int intel_modeset_vga_set_state(struct drm_i915_private *i915, bool state)
> >   {
> > -	unsigned reg = INTEL_GEN(dev_priv) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL;
> > +	unsigned reg = INTEL_GEN(i915) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL;
> >   	u16 gmch_ctrl;
> > -	if (pci_read_config_word(dev_priv->bridge_dev, reg, &gmch_ctrl)) {
> > +	if (pci_read_config_word(i915->bridge_dev, reg, &gmch_ctrl)) {
> >   		DRM_ERROR("failed to read control word\n");
> >   		return -EIO;
> >   	}
> > @@ -16926,7 +16926,7 @@ int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state)
> >   	else
> >   		gmch_ctrl |= INTEL_GMCH_VGA_DISABLE;
> > -	if (pci_write_config_word(dev_priv->bridge_dev, reg, gmch_ctrl)) {
> > +	if (pci_write_config_word(i915->bridge_dev, reg, gmch_ctrl)) {
> >   		DRM_ERROR("failed to write control word\n");
> >   		return -EIO;
> >   	}
> > @@ -16980,7 +16980,7 @@ struct intel_display_error_state {
> >   };
> >   struct intel_display_error_state *
> > -intel_display_capture_error_state(struct drm_i915_private *dev_priv)
> > +intel_display_capture_error_state(struct drm_i915_private *i915)
> >   {
> >   	struct intel_display_error_state *error;
> >   	int transcoders[] = {
> > @@ -16993,19 +16993,19 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv)
> >   	BUILD_BUG_ON(ARRAY_SIZE(transcoders) != ARRAY_SIZE(error->transcoder));
> > -	if (!HAS_DISPLAY(dev_priv))
> > +	if (!HAS_DISPLAY(i915))
> >   		return NULL;
> >   	error = kzalloc(sizeof(*error), GFP_ATOMIC);
> >   	if (error == NULL)
> >   		return NULL;
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		error->power_well_driver = I915_READ(HSW_PWR_WELL_CTL2);
> > -	for_each_pipe(dev_priv, i) {
> > +	for_each_pipe(i915, i) {
> >   		error->pipe[i].power_domain_on =
> > -			__intel_display_power_is_enabled(dev_priv,
> > +			__intel_display_power_is_enabled(i915,
> >   							 POWER_DOMAIN_PIPE(i));
> >   		if (!error->pipe[i].power_domain_on)
> >   			continue;
> > @@ -17016,32 +17016,32 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv)
> >   		error->plane[i].control = I915_READ(DSPCNTR(i));
> >   		error->plane[i].stride = I915_READ(DSPSTRIDE(i));
> > -		if (INTEL_GEN(dev_priv) <= 3) {
> > +		if (INTEL_GEN(i915) <= 3) {
> >   			error->plane[i].size = I915_READ(DSPSIZE(i));
> >   			error->plane[i].pos = I915_READ(DSPPOS(i));
> >   		}
> > -		if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv))
> > +		if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915))
> >   			error->plane[i].addr = I915_READ(DSPADDR(i));
> > -		if (INTEL_GEN(dev_priv) >= 4) {
> > +		if (INTEL_GEN(i915) >= 4) {
> >   			error->plane[i].surface = I915_READ(DSPSURF(i));
> >   			error->plane[i].tile_offset = I915_READ(DSPTILEOFF(i));
> >   		}
> >   		error->pipe[i].source = I915_READ(PIPESRC(i));
> > -		if (HAS_GMCH(dev_priv))
> > +		if (HAS_GMCH(i915))
> >   			error->pipe[i].stat = I915_READ(PIPESTAT(i));
> >   	}
> >   	for (i = 0; i < ARRAY_SIZE(error->transcoder); i++) {
> >   		enum transcoder cpu_transcoder = transcoders[i];
> > -		if (!INTEL_INFO(dev_priv)->trans_offsets[cpu_transcoder])
> > +		if (!INTEL_INFO(i915)->trans_offsets[cpu_transcoder])
> >   			continue;
> >   		error->transcoder[i].available = true;
> >   		error->transcoder[i].power_domain_on =
> > -			__intel_display_power_is_enabled(dev_priv,
> > +			__intel_display_power_is_enabled(i915,
> >   				POWER_DOMAIN_TRANSCODER(cpu_transcoder));
> >   		if (!error->transcoder[i].power_domain_on)
> >   			continue;
> > @@ -17066,17 +17066,17 @@ void
> >   intel_display_print_error_state(struct drm_i915_error_state_buf *m,
> >   				struct intel_display_error_state *error)
> >   {
> > -	struct drm_i915_private *dev_priv = m->i915;
> > +	struct drm_i915_private *i915 = m->i915;
> >   	int i;
> >   	if (!error)
> >   		return;
> > -	err_printf(m, "Num Pipes: %d\n", INTEL_INFO(dev_priv)->num_pipes);
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	err_printf(m, "Num Pipes: %d\n", INTEL_INFO(i915)->num_pipes);
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		err_printf(m, "PWR_WELL_CTL2: %08x\n",
> >   			   error->power_well_driver);
> > -	for_each_pipe(dev_priv, i) {
> > +	for_each_pipe(i915, i) {
> >   		err_printf(m, "Pipe [%d]:\n", i);
> >   		err_printf(m, "  Power: %s\n",
> >   			   onoff(error->pipe[i].power_domain_on));
> > @@ -17086,13 +17086,13 @@ intel_display_print_error_state(struct drm_i915_error_state_buf *m,
> >   		err_printf(m, "Plane [%d]:\n", i);
> >   		err_printf(m, "  CNTR: %08x\n", error->plane[i].control);
> >   		err_printf(m, "  STRIDE: %08x\n", error->plane[i].stride);
> > -		if (INTEL_GEN(dev_priv) <= 3) {
> > +		if (INTEL_GEN(i915) <= 3) {
> >   			err_printf(m, "  SIZE: %08x\n", error->plane[i].size);
> >   			err_printf(m, "  POS: %08x\n", error->plane[i].pos);
> >   		}
> > -		if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv))
> > +		if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915))
> >   			err_printf(m, "  ADDR: %08x\n", error->plane[i].addr);
> > -		if (INTEL_GEN(dev_priv) >= 4) {
> > +		if (INTEL_GEN(i915) >= 4) {
> >   			err_printf(m, "  SURF: %08x\n", error->plane[i].surface);
> >   			err_printf(m, "  TILEOFF: %08x\n", error->plane[i].tile_offset);
> >   		}
> > diff --git a/drivers/gpu/drm/i915/intel_display.h b/drivers/gpu/drm/i915/intel_display.h
> > index ee6b8194a459..7e4b4b1c0ac6 100644
> > --- a/drivers/gpu/drm/i915/intel_display.h
> > +++ b/drivers/gpu/drm/i915/intel_display.h
> > @@ -125,7 +125,7 @@ enum i9xx_plane_id {
> >   };
> >   #define plane_name(p) ((p) + 'A')
> > -#define sprite_name(p, s) ((p) * RUNTIME_INFO(dev_priv)->num_sprites[(p)] + (s) + 'A')
> > +#define sprite_name(p, s) ((p) * RUNTIME_INFO(i915)->num_sprites[(p)] + (s) + 'A')
> >   /*
> >    * Per-pipe plane identifier.
> > @@ -353,8 +353,8 @@ void intel_link_compute_m_n(u16 bpp, int nlanes,
> >   			    struct intel_link_m_n *m_n,
> >   			    bool constant_n);
> >   bool is_ccs_modifier(u64 modifier);
> > -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv);
> > -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv,
> > +void lpt_disable_clkout_dp(struct drm_i915_private *i915);
> > +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915,
> >   			      u32 pixel_format, u64 modifier);
> >   bool intel_plane_can_remap(const struct intel_plane_state *plane_state);
> > diff --git a/drivers/gpu/drm/i915/intel_display_power.c b/drivers/gpu/drm/i915/intel_display_power.c
> > index bb9ef1cea5db..9acfad5c83ef 100644
> > --- a/drivers/gpu/drm/i915/intel_display_power.c
> > +++ b/drivers/gpu/drm/i915/intel_display_power.c
> > @@ -17,7 +17,7 @@
> >   #include "intel_hotplug.h"
> >   #include "intel_sideband.h"
> > -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915,
> >   					 enum i915_power_well_id power_well_id);
> >   const char *
> > @@ -122,42 +122,42 @@ intel_display_power_domain_str(enum intel_display_power_domain domain)
> >   	}
> >   }
> > -static void intel_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void intel_power_well_enable(struct drm_i915_private *i915,
> >   				    struct i915_power_well *power_well)
> >   {
> >   	DRM_DEBUG_KMS("enabling %s\n", power_well->desc->name);
> > -	power_well->desc->ops->enable(dev_priv, power_well);
> > +	power_well->desc->ops->enable(i915, power_well);
> >   	power_well->hw_enabled = true;
> >   }
> > -static void intel_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void intel_power_well_disable(struct drm_i915_private *i915,
> >   				     struct i915_power_well *power_well)
> >   {
> >   	DRM_DEBUG_KMS("disabling %s\n", power_well->desc->name);
> >   	power_well->hw_enabled = false;
> > -	power_well->desc->ops->disable(dev_priv, power_well);
> > +	power_well->desc->ops->disable(i915, power_well);
> >   }
> > -static void intel_power_well_get(struct drm_i915_private *dev_priv,
> > +static void intel_power_well_get(struct drm_i915_private *i915,
> >   				 struct i915_power_well *power_well)
> >   {
> >   	if (!power_well->count++)
> > -		intel_power_well_enable(dev_priv, power_well);
> > +		intel_power_well_enable(i915, power_well);
> >   }
> > -static void intel_power_well_put(struct drm_i915_private *dev_priv,
> > +static void intel_power_well_put(struct drm_i915_private *i915,
> >   				 struct i915_power_well *power_well)
> >   {
> >   	WARN(!power_well->count, "Use count on power well %s is already zero",
> >   	     power_well->desc->name);
> >   	if (!--power_well->count)
> > -		intel_power_well_disable(dev_priv, power_well);
> > +		intel_power_well_disable(i915, power_well);
> >   }
> >   /**
> >    * __intel_display_power_is_enabled - unlocked check for a power domain
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to check
> >    *
> >    * This is the unlocked version of intel_display_power_is_enabled() and should
> > @@ -167,18 +167,18 @@ static void intel_power_well_put(struct drm_i915_private *dev_priv,
> >    * Returns:
> >    * True when the power domain is enabled, false otherwise.
> >    */
> > -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> > +bool __intel_display_power_is_enabled(struct drm_i915_private *i915,
> >   				      enum intel_display_power_domain domain)
> >   {
> >   	struct i915_power_well *power_well;
> >   	bool is_enabled;
> > -	if (dev_priv->runtime_pm.suspended)
> > +	if (i915->runtime_pm.suspended)
> >   		return false;
> >   	is_enabled = true;
> > -	for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain)) {
> > +	for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain)) {
> >   		if (power_well->desc->always_on)
> >   			continue;
> > @@ -193,7 +193,7 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_display_power_is_enabled - check for a power domain
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to check
> >    *
> >    * This function can be used to check the hw power domain state. It is mostly
> > @@ -208,16 +208,16 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> >    * Returns:
> >    * True when the power domain is enabled, false otherwise.
> >    */
> > -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_display_power_is_enabled(struct drm_i915_private *i915,
> >   				    enum intel_display_power_domain domain)
> >   {
> >   	struct i915_power_domains *power_domains;
> >   	bool ret;
> > -	power_domains = &dev_priv->power_domains;
> > +	power_domains = &i915->power_domains;
> >   	mutex_lock(&power_domains->lock);
> > -	ret = __intel_display_power_is_enabled(dev_priv, domain);
> > +	ret = __intel_display_power_is_enabled(i915, domain);
> >   	mutex_unlock(&power_domains->lock);
> >   	return ret;
> > @@ -229,10 +229,10 @@ bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> >    * to be enabled, and it will only be disabled if none of the registers is
> >    * requesting it to be enabled.
> >    */
> > -static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_power_well_post_enable(struct drm_i915_private *i915,
> >   				       u8 irq_pipe_mask, bool has_vga)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	/*
> >   	 * After we re-enable the power well, if we touch VGA register 0x3d5
> > @@ -251,31 +251,31 @@ static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv,
> >   	}
> >   	if (irq_pipe_mask)
> > -		gen8_irq_power_well_post_enable(dev_priv, irq_pipe_mask);
> > +		gen8_irq_power_well_post_enable(i915, irq_pipe_mask);
> >   }
> > -static void hsw_power_well_pre_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_power_well_pre_disable(struct drm_i915_private *i915,
> >   				       u8 irq_pipe_mask)
> >   {
> >   	if (irq_pipe_mask)
> > -		gen8_irq_power_well_pre_disable(dev_priv, irq_pipe_mask);
> > +		gen8_irq_power_well_pre_disable(i915, irq_pipe_mask);
> >   }
> > -static void hsw_wait_for_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_wait_for_power_well_enable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> >   	int pw_idx = power_well->desc->hsw.idx;
> >   	/* Timeout for PW1:10 us, AUX:not specified, other PWs:20 us. */
> > -	WARN_ON(intel_wait_for_register(&dev_priv->uncore,
> > +	WARN_ON(intel_wait_for_register(&i915->uncore,
> >   					regs->driver,
> >   					HSW_PWR_WELL_CTL_STATE(pw_idx),
> >   					HSW_PWR_WELL_CTL_STATE(pw_idx),
> >   					1));
> >   }
> > -static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv,
> > +static u32 hsw_power_well_requesters(struct drm_i915_private *i915,
> >   				     const struct i915_power_well_regs *regs,
> >   				     int pw_idx)
> >   {
> > @@ -291,7 +291,7 @@ static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv,
> >   	return ret;
> >   }
> > -static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_wait_for_power_well_disable(struct drm_i915_private *i915,
> >   					    struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -310,7 +310,7 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> >   	 */
> >   	wait_for((disabled = !(I915_READ(regs->driver) &
> >   			       HSW_PWR_WELL_CTL_STATE(pw_idx))) ||
> > -		 (reqs = hsw_power_well_requesters(dev_priv, regs, pw_idx)), 1);
> > +		 (reqs = hsw_power_well_requesters(i915, regs, pw_idx)), 1);
> >   	if (disabled)
> >   		return;
> > @@ -319,16 +319,16 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
> >   		      !!(reqs & 1), !!(reqs & 2), !!(reqs & 4), !!(reqs & 8));
> >   }
> > -static void gen9_wait_for_power_well_fuses(struct drm_i915_private *dev_priv,
> > +static void gen9_wait_for_power_well_fuses(struct drm_i915_private *i915,
> >   					   enum skl_power_gate pg)
> >   {
> >   	/* Timeout 5us for PG#0, for other PGs 1us */
> > -	WARN_ON(intel_wait_for_register(&dev_priv->uncore, SKL_FUSE_STATUS,
> > +	WARN_ON(intel_wait_for_register(&i915->uncore, SKL_FUSE_STATUS,
> >   					SKL_FUSE_PG_DIST_STATUS(pg),
> >   					SKL_FUSE_PG_DIST_STATUS(pg), 1));
> >   }
> > -static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_power_well_enable(struct drm_i915_private *i915,
> >   				  struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -338,7 +338,7 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> >   	u32 val;
> >   	if (wait_fuses) {
> > -		pg = INTEL_GEN(dev_priv) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) :
> > +		pg = INTEL_GEN(i915) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) :
> >   						 SKL_PW_CTL_IDX_TO_PG(pw_idx);
> >   		/*
> >   		 * For PW1 we have to wait both for the PW0/PG0 fuse state
> > @@ -348,15 +348,15 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> >   		 * after the enabling.
> >   		 */
> >   		if (pg == SKL_PG1)
> > -			gen9_wait_for_power_well_fuses(dev_priv, SKL_PG0);
> > +			gen9_wait_for_power_well_fuses(i915, SKL_PG0);
> >   	}
> >   	val = I915_READ(regs->driver);
> >   	I915_WRITE(regs->driver, val | HSW_PWR_WELL_CTL_REQ(pw_idx));
> > -	hsw_wait_for_power_well_enable(dev_priv, power_well);
> > +	hsw_wait_for_power_well_enable(i915, power_well);
> >   	/* Display WA #1178: cnl */
> > -	if (IS_CANNONLAKE(dev_priv) &&
> > +	if (IS_CANNONLAKE(i915) &&
> >   	    pw_idx >= GLK_PW_CTL_IDX_AUX_B &&
> >   	    pw_idx <= CNL_PW_CTL_IDX_AUX_F) {
> >   		val = I915_READ(CNL_AUX_ANAOVRD1(pw_idx));
> > @@ -365,32 +365,32 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
> >   	}
> >   	if (wait_fuses)
> > -		gen9_wait_for_power_well_fuses(dev_priv, pg);
> > +		gen9_wait_for_power_well_fuses(i915, pg);
> > -	hsw_power_well_post_enable(dev_priv,
> > +	hsw_power_well_post_enable(i915,
> >   				   power_well->desc->hsw.irq_pipe_mask,
> >   				   power_well->desc->hsw.has_vga);
> >   }
> > -static void hsw_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_power_well_disable(struct drm_i915_private *i915,
> >   				   struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> >   	int pw_idx = power_well->desc->hsw.idx;
> >   	u32 val;
> > -	hsw_power_well_pre_disable(dev_priv,
> > +	hsw_power_well_pre_disable(i915,
> >   				   power_well->desc->hsw.irq_pipe_mask);
> >   	val = I915_READ(regs->driver);
> >   	I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
> > -	hsw_wait_for_power_well_disable(dev_priv, power_well);
> > +	hsw_wait_for_power_well_disable(i915, power_well);
> >   }
> >   #define ICL_AUX_PW_TO_PORT(pw_idx)	((pw_idx) - ICL_PW_CTL_IDX_AUX_A)
> >   static void
> > -icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> > +icl_combo_phy_aux_power_well_enable(struct drm_i915_private *i915,
> >   				    struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -404,12 +404,12 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(ICL_PORT_CL_DW12(port));
> >   	I915_WRITE(ICL_PORT_CL_DW12(port), val | ICL_LANE_ENABLE_AUX);
> > -	hsw_wait_for_power_well_enable(dev_priv, power_well);
> > +	hsw_wait_for_power_well_enable(i915, power_well);
> >   	/* Display WA #1178: icl */
> > -	if (IS_ICELAKE(dev_priv) &&
> > +	if (IS_ICELAKE(i915) &&
> >   	    pw_idx >= ICL_PW_CTL_IDX_AUX_A && pw_idx <= ICL_PW_CTL_IDX_AUX_B &&
> > -	    !intel_bios_is_port_edp(dev_priv, port)) {
> > +	    !intel_bios_is_port_edp(i915, port)) {
> >   		val = I915_READ(ICL_AUX_ANAOVRD1(pw_idx));
> >   		val |= ICL_AUX_ANAOVRD1_ENABLE | ICL_AUX_ANAOVRD1_LDO_BYPASS;
> >   		I915_WRITE(ICL_AUX_ANAOVRD1(pw_idx), val);
> > @@ -417,7 +417,7 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv,
> > +icl_combo_phy_aux_power_well_disable(struct drm_i915_private *i915,
> >   				     struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -431,14 +431,14 @@ icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(regs->driver);
> >   	I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
> > -	hsw_wait_for_power_well_disable(dev_priv, power_well);
> > +	hsw_wait_for_power_well_disable(i915, power_well);
> >   }
> >   #define ICL_AUX_PW_TO_CH(pw_idx)	\
> >   	((pw_idx) - ICL_PW_CTL_IDX_AUX_A + AUX_CH_A)
> >   static void
> > -icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> > +icl_tc_phy_aux_power_well_enable(struct drm_i915_private *i915,
> >   				 struct i915_power_well *power_well)
> >   {
> >   	enum aux_ch aux_ch = ICL_AUX_PW_TO_CH(power_well->desc->hsw.idx);
> > @@ -450,7 +450,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> >   		val |= DP_AUX_CH_CTL_TBT_IO;
> >   	I915_WRITE(DP_AUX_CH_CTL(aux_ch), val);
> > -	hsw_power_well_enable(dev_priv, power_well);
> > +	hsw_power_well_enable(i915, power_well);
> >   }
> >   /*
> > @@ -458,7 +458,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
> >    * enable it, so check if it's enabled and also check if we've requested it to
> >    * be enabled.
> >    */
> > -static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool hsw_power_well_enabled(struct drm_i915_private *i915,
> >   				   struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -476,14 +476,14 @@ static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv,
> >   	 * BIOS's own request bits, which are forced-on for these power wells
> >   	 * when exiting DC5/6.
> >   	 */
> > -	if (IS_GEN(dev_priv, 9) && !IS_GEN9_LP(dev_priv) &&
> > +	if (IS_GEN(i915, 9) && !IS_GEN9_LP(i915) &&
> >   	    (id == SKL_DISP_PW_1 || id == SKL_DISP_PW_MISC_IO))
> >   		val |= I915_READ(regs->bios);
> >   	return (val & mask) == mask;
> >   }
> > -static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> > +static void assert_can_enable_dc9(struct drm_i915_private *i915)
> >   {
> >   	WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_DC9),
> >   		  "DC9 already programmed to be enabled.\n");
> > @@ -492,7 +492,7 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> >   	WARN_ONCE(I915_READ(HSW_PWR_WELL_CTL2) &
> >   		  HSW_PWR_WELL_CTL_REQ(SKL_PW_CTL_IDX_PW_2),
> >   		  "Power well 2 on.\n");
> > -	WARN_ONCE(intel_irqs_enabled(dev_priv),
> > +	WARN_ONCE(intel_irqs_enabled(i915),
> >   		  "Interrupts not disabled yet.\n");
> >   	 /*
> > @@ -504,9 +504,9 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
> >   	  */
> >   }
> > -static void assert_can_disable_dc9(struct drm_i915_private *dev_priv)
> > +static void assert_can_disable_dc9(struct drm_i915_private *i915)
> >   {
> > -	WARN_ONCE(intel_irqs_enabled(dev_priv),
> > +	WARN_ONCE(intel_irqs_enabled(i915),
> >   		  "Interrupts not disabled yet.\n");
> >   	WARN_ONCE(I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5,
> >   		  "DC5 still not disabled.\n");
> > @@ -520,7 +520,7 @@ static void assert_can_disable_dc9(struct drm_i915_private *dev_priv)
> >   	  */
> >   }
> > -static void gen9_write_dc_state(struct drm_i915_private *dev_priv,
> > +static void gen9_write_dc_state(struct drm_i915_private *i915,
> >   				u32 state)
> >   {
> >   	int rewrites = 0;
> > @@ -557,14 +557,14 @@ static void gen9_write_dc_state(struct drm_i915_private *dev_priv,
> >   			      state, rewrites);
> >   }
> > -static u32 gen9_dc_mask(struct drm_i915_private *dev_priv)
> > +static u32 gen9_dc_mask(struct drm_i915_private *i915)
> >   {
> >   	u32 mask;
> >   	mask = DC_STATE_EN_UPTO_DC5;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		mask |= DC_STATE_EN_UPTO_DC6 | DC_STATE_EN_DC9;
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		mask |= DC_STATE_EN_DC9;
> >   	else
> >   		mask |= DC_STATE_EN_UPTO_DC6;
> > @@ -572,20 +572,20 @@ static u32 gen9_dc_mask(struct drm_i915_private *dev_priv)
> >   	return mask;
> >   }
> > -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv)
> > +void gen9_sanitize_dc_state(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	val = I915_READ(DC_STATE_EN) & gen9_dc_mask(dev_priv);
> > +	val = I915_READ(DC_STATE_EN) & gen9_dc_mask(i915);
> >   	DRM_DEBUG_KMS("Resetting DC state tracking from %02x to %02x\n",
> > -		      dev_priv->csr.dc_state, val);
> > -	dev_priv->csr.dc_state = val;
> > +		      i915->csr.dc_state, val);
> > +	i915->csr.dc_state = val;
> >   }
> >   /**
> >    * gen9_set_dc_state - set target display C power state
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @state: target DC power state
> >    * - DC_STATE_DISABLE
> >    * - DC_STATE_EN_UPTO_DC5
> > @@ -606,35 +606,35 @@ void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv)
> >    * back on and register state is restored. This is guaranteed by the MMIO write
> >    * to DC_STATE_EN blocking until the state is restored.
> >    */
> > -static void gen9_set_dc_state(struct drm_i915_private *dev_priv, u32 state)
> > +static void gen9_set_dc_state(struct drm_i915_private *i915, u32 state)
> >   {
> >   	u32 val;
> >   	u32 mask;
> > -	if (WARN_ON_ONCE(state & ~dev_priv->csr.allowed_dc_mask))
> > -		state &= dev_priv->csr.allowed_dc_mask;
> > +	if (WARN_ON_ONCE(state & ~i915->csr.allowed_dc_mask))
> > +		state &= i915->csr.allowed_dc_mask;
> >   	val = I915_READ(DC_STATE_EN);
> > -	mask = gen9_dc_mask(dev_priv);
> > +	mask = gen9_dc_mask(i915);
> >   	DRM_DEBUG_KMS("Setting DC state from %02x to %02x\n",
> >   		      val & mask, state);
> >   	/* Check if DMC is ignoring our DC state requests */
> > -	if ((val & mask) != dev_priv->csr.dc_state)
> > +	if ((val & mask) != i915->csr.dc_state)
> >   		DRM_ERROR("DC state mismatch (0x%x -> 0x%x)\n",
> > -			  dev_priv->csr.dc_state, val & mask);
> > +			  i915->csr.dc_state, val & mask);
> >   	val &= ~mask;
> >   	val |= state;
> > -	gen9_write_dc_state(dev_priv, val);
> > +	gen9_write_dc_state(i915, val);
> > -	dev_priv->csr.dc_state = val & mask;
> > +	i915->csr.dc_state = val & mask;
> >   }
> > -void bxt_enable_dc9(struct drm_i915_private *dev_priv)
> > +void bxt_enable_dc9(struct drm_i915_private *i915)
> >   {
> > -	assert_can_enable_dc9(dev_priv);
> > +	assert_can_enable_dc9(i915);
> >   	DRM_DEBUG_KMS("Enabling DC9\n");
> >   	/*
> > @@ -642,23 +642,23 @@ void bxt_enable_dc9(struct drm_i915_private *dev_priv)
> >   	 * platforms with South Display Engine on PCH,
> >   	 * because PPS registers are always on.
> >   	 */
> > -	if (!HAS_PCH_SPLIT(dev_priv))
> > -		intel_power_sequencer_reset(dev_priv);
> > -	gen9_set_dc_state(dev_priv, DC_STATE_EN_DC9);
> > +	if (!HAS_PCH_SPLIT(i915))
> > +		intel_power_sequencer_reset(i915);
> > +	gen9_set_dc_state(i915, DC_STATE_EN_DC9);
> >   }
> > -void bxt_disable_dc9(struct drm_i915_private *dev_priv)
> > +void bxt_disable_dc9(struct drm_i915_private *i915)
> >   {
> > -	assert_can_disable_dc9(dev_priv);
> > +	assert_can_disable_dc9(i915);
> >   	DRM_DEBUG_KMS("Disabling DC9\n");
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> > -	intel_pps_unlock_regs_wa(dev_priv);
> > +	intel_pps_unlock_regs_wa(i915);
> >   }
> > -static void assert_csr_loaded(struct drm_i915_private *dev_priv)
> > +static void assert_csr_loaded(struct drm_i915_private *i915)
> >   {
> >   	WARN_ONCE(!I915_READ(CSR_PROGRAM(0)),
> >   		  "CSR program storage start is NULL\n");
> > @@ -667,12 +667,12 @@ static void assert_csr_loaded(struct drm_i915_private *dev_priv)
> >   }
> >   static struct i915_power_well *
> > -lookup_power_well(struct drm_i915_private *dev_priv,
> > +lookup_power_well(struct drm_i915_private *i915,
> >   		  enum i915_power_well_id power_well_id)
> >   {
> >   	struct i915_power_well *power_well;
> > -	for_each_power_well(dev_priv, power_well)
> > +	for_each_power_well(i915, power_well)
> >   		if (power_well->desc->id == power_well_id)
> >   			return power_well;
> > @@ -684,62 +684,62 @@ lookup_power_well(struct drm_i915_private *dev_priv,
> >   	 * our driver.
> >   	 */
> >   	WARN(1, "Power well %d not defined for this platform\n", power_well_id);
> > -	return &dev_priv->power_domains.power_wells[0];
> > +	return &i915->power_domains.power_wells[0];
> >   }
> > -static void assert_can_enable_dc5(struct drm_i915_private *dev_priv)
> > +static void assert_can_enable_dc5(struct drm_i915_private *i915)
> >   {
> > -	bool pg2_enabled = intel_display_power_well_is_enabled(dev_priv,
> > +	bool pg2_enabled = intel_display_power_well_is_enabled(i915,
> >   					SKL_DISP_PW_2);
> >   	WARN_ONCE(pg2_enabled, "PG2 not disabled to enable DC5.\n");
> >   	WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5),
> >   		  "DC5 already programmed to be enabled.\n");
> > -	assert_rpm_wakelock_held(dev_priv);
> > +	assert_rpm_wakelock_held(i915);
> > -	assert_csr_loaded(dev_priv);
> > +	assert_csr_loaded(i915);
> >   }
> > -void gen9_enable_dc5(struct drm_i915_private *dev_priv)
> > +void gen9_enable_dc5(struct drm_i915_private *i915)
> >   {
> > -	assert_can_enable_dc5(dev_priv);
> > +	assert_can_enable_dc5(i915);
> >   	DRM_DEBUG_KMS("Enabling DC5\n");
> >   	/* Wa Display #1183: skl,kbl,cfl */
> > -	if (IS_GEN9_BC(dev_priv))
> > +	if (IS_GEN9_BC(i915))
> >   		I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
> >   			   SKL_SELECT_ALTERNATE_DC_EXIT);
> > -	gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC5);
> > +	gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC5);
> >   }
> > -static void assert_can_enable_dc6(struct drm_i915_private *dev_priv)
> > +static void assert_can_enable_dc6(struct drm_i915_private *i915)
> >   {
> >   	WARN_ONCE(I915_READ(UTIL_PIN_CTL) & UTIL_PIN_ENABLE,
> >   		  "Backlight is not disabled.\n");
> >   	WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC6),
> >   		  "DC6 already programmed to be enabled.\n");
> > -	assert_csr_loaded(dev_priv);
> > +	assert_csr_loaded(i915);
> >   }
> > -void skl_enable_dc6(struct drm_i915_private *dev_priv)
> > +void skl_enable_dc6(struct drm_i915_private *i915)
> >   {
> > -	assert_can_enable_dc6(dev_priv);
> > +	assert_can_enable_dc6(i915);
> >   	DRM_DEBUG_KMS("Enabling DC6\n");
> >   	/* Wa Display #1183: skl,kbl,cfl */
> > -	if (IS_GEN9_BC(dev_priv))
> > +	if (IS_GEN9_BC(i915))
> >   		I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
> >   			   SKL_SELECT_ALTERNATE_DC_EXIT);
> > -	gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC6);
> > +	gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC6);
> >   }
> > -static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv,
> > +static void hsw_power_well_sync_hw(struct drm_i915_private *i915,
> >   				   struct i915_power_well *power_well)
> >   {
> >   	const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
> > @@ -757,52 +757,52 @@ static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> > -	bxt_ddi_phy_init(dev_priv, power_well->desc->bxt.phy);
> > +	bxt_ddi_phy_init(i915, power_well->desc->bxt.phy);
> >   }
> > -static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> >   					    struct i915_power_well *power_well)
> >   {
> > -	bxt_ddi_phy_uninit(dev_priv, power_well->desc->bxt.phy);
> > +	bxt_ddi_phy_uninit(i915, power_well->desc->bxt.phy);
> >   }
> > -static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *i915,
> >   					    struct i915_power_well *power_well)
> >   {
> > -	return bxt_ddi_phy_is_enabled(dev_priv, power_well->desc->bxt.phy);
> > +	return bxt_ddi_phy_is_enabled(i915, power_well->desc->bxt.phy);
> >   }
> > -static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *dev_priv)
> > +static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *i915)
> >   {
> >   	struct i915_power_well *power_well;
> > -	power_well = lookup_power_well(dev_priv, BXT_DISP_PW_DPIO_CMN_A);
> > +	power_well = lookup_power_well(i915, BXT_DISP_PW_DPIO_CMN_A);
> >   	if (power_well->count > 0)
> > -		bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
> > +		bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy);
> > -	power_well = lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> > +	power_well = lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> >   	if (power_well->count > 0)
> > -		bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
> > +		bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy);
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > -		power_well = lookup_power_well(dev_priv,
> > +	if (IS_GEMINILAKE(i915)) {
> > +		power_well = lookup_power_well(i915,
> >   					       GLK_DISP_PW_DPIO_CMN_C);
> >   		if (power_well->count > 0)
> > -			bxt_ddi_phy_verify_state(dev_priv,
> > +			bxt_ddi_phy_verify_state(i915,
> >   						 power_well->desc->bxt.phy);
> >   	}
> >   }
> > -static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> >   	return (I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5_DC6_MASK) == 0;
> >   }
> > -static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv)
> > +static void gen9_assert_dbuf_enabled(struct drm_i915_private *i915)
> >   {
> >   	u32 tmp = I915_READ(DBUF_CTL);
> > @@ -811,92 +811,92 @@ static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv)
> >   	     "Unexpected DBuf power power state (0x%08x)\n", tmp);
> >   }
> > -static void gen9_dc_off_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void gen9_dc_off_power_well_enable(struct drm_i915_private *i915,
> >   					  struct i915_power_well *power_well)
> >   {
> >   	struct intel_cdclk_state cdclk_state = {};
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> > -	dev_priv->display.get_cdclk(dev_priv, &cdclk_state);
> > +	i915->display.get_cdclk(i915, &cdclk_state);
> >   	/* Can't read out voltage_level so can't use intel_cdclk_changed() */
> > -	WARN_ON(intel_cdclk_needs_modeset(&dev_priv->cdclk.hw, &cdclk_state));
> > +	WARN_ON(intel_cdclk_needs_modeset(&i915->cdclk.hw, &cdclk_state));
> > -	gen9_assert_dbuf_enabled(dev_priv);
> > +	gen9_assert_dbuf_enabled(i915);
> > -	if (IS_GEN9_LP(dev_priv))
> > -		bxt_verify_ddi_phy_power_wells(dev_priv);
> > +	if (IS_GEN9_LP(i915))
> > +		bxt_verify_ddi_phy_power_wells(i915);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		/*
> >   		 * DMC retains HW context only for port A, the other combo
> >   		 * PHY's HW context for port B is lost after DC transitions,
> >   		 * so we need to restore it manually.
> >   		 */
> > -		intel_combo_phy_init(dev_priv);
> > +		intel_combo_phy_init(i915);
> >   }
> > -static void gen9_dc_off_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void gen9_dc_off_power_well_disable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> > -	if (!dev_priv->csr.dmc_payload)
> > +	if (!i915->csr.dmc_payload)
> >   		return;
> > -	if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6)
> > -		skl_enable_dc6(dev_priv);
> > -	else if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)
> > -		gen9_enable_dc5(dev_priv);
> > +	if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6)
> > +		skl_enable_dc6(i915);
> > +	else if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)
> > +		gen9_enable_dc5(i915);
> >   }
> > -static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *dev_priv,
> > +static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *i915,
> >   					 struct i915_power_well *power_well)
> >   {
> >   }
> > -static void i9xx_always_on_power_well_noop(struct drm_i915_private *dev_priv,
> > +static void i9xx_always_on_power_well_noop(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> >   }
> > -static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *i915,
> >   					     struct i915_power_well *power_well)
> >   {
> >   	return true;
> >   }
> > -static void i830_pipes_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void i830_pipes_power_well_enable(struct drm_i915_private *i915,
> >   					 struct i915_power_well *power_well)
> >   {
> >   	if ((I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE) == 0)
> > -		i830_enable_pipe(dev_priv, PIPE_A);
> > +		i830_enable_pipe(i915, PIPE_A);
> >   	if ((I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE) == 0)
> > -		i830_enable_pipe(dev_priv, PIPE_B);
> > +		i830_enable_pipe(i915, PIPE_B);
> >   }
> > -static void i830_pipes_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void i830_pipes_power_well_disable(struct drm_i915_private *i915,
> >   					  struct i915_power_well *power_well)
> >   {
> > -	i830_disable_pipe(dev_priv, PIPE_B);
> > -	i830_disable_pipe(dev_priv, PIPE_A);
> > +	i830_disable_pipe(i915, PIPE_B);
> > +	i830_disable_pipe(i915, PIPE_A);
> >   }
> > -static bool i830_pipes_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool i830_pipes_power_well_enabled(struct drm_i915_private *i915,
> >   					  struct i915_power_well *power_well)
> >   {
> >   	return I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE &&
> >   		I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE;
> >   }
> > -static void i830_pipes_power_well_sync_hw(struct drm_i915_private *dev_priv,
> > +static void i830_pipes_power_well_sync_hw(struct drm_i915_private *i915,
> >   					  struct i915_power_well *power_well)
> >   {
> >   	if (power_well->count > 0)
> > -		i830_pipes_power_well_enable(dev_priv, power_well);
> > +		i830_pipes_power_well_enable(i915, power_well);
> >   	else
> > -		i830_pipes_power_well_disable(dev_priv, power_well);
> > +		i830_pipes_power_well_disable(i915, power_well);
> >   }
> > -static void vlv_set_power_well(struct drm_i915_private *dev_priv,
> > +static void vlv_set_power_well(struct drm_i915_private *i915,
> >   			       struct i915_power_well *power_well, bool enable)
> >   {
> >   	int pw_idx = power_well->desc->vlv.idx;
> > @@ -908,43 +908,43 @@ static void vlv_set_power_well(struct drm_i915_private *dev_priv,
> >   	state = enable ? PUNIT_PWRGT_PWR_ON(pw_idx) :
> >   			 PUNIT_PWRGT_PWR_GATE(pw_idx);
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> >   #define COND \
> > -	((vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask) == state)
> > +	((vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask) == state)
> >   	if (COND)
> >   		goto out;
> > -	ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL);
> > +	ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL);
> >   	ctrl &= ~mask;
> >   	ctrl |= state;
> > -	vlv_punit_write(dev_priv, PUNIT_REG_PWRGT_CTRL, ctrl);
> > +	vlv_punit_write(i915, PUNIT_REG_PWRGT_CTRL, ctrl);
> >   	if (wait_for(COND, 100))
> >   		DRM_ERROR("timeout setting power well state %08x (%08x)\n",
> >   			  state,
> > -			  vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL));
> > +			  vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL));
> >   #undef COND
> >   out:
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   }
> > -static void vlv_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void vlv_power_well_enable(struct drm_i915_private *i915,
> >   				  struct i915_power_well *power_well)
> >   {
> > -	vlv_set_power_well(dev_priv, power_well, true);
> > +	vlv_set_power_well(i915, power_well, true);
> >   }
> > -static void vlv_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void vlv_power_well_disable(struct drm_i915_private *i915,
> >   				   struct i915_power_well *power_well)
> >   {
> > -	vlv_set_power_well(dev_priv, power_well, false);
> > +	vlv_set_power_well(i915, power_well, false);
> >   }
> > -static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool vlv_power_well_enabled(struct drm_i915_private *i915,
> >   				   struct i915_power_well *power_well)
> >   {
> >   	int pw_idx = power_well->desc->vlv.idx;
> > @@ -956,9 +956,9 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> >   	mask = PUNIT_PWRGT_MASK(pw_idx);
> >   	ctrl = PUNIT_PWRGT_PWR_ON(pw_idx);
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> > -	state = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask;
> > +	state = vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask;
> >   	/*
> >   	 * We only ever set the power-on and power-gate states, anything
> >   	 * else is unexpected.
> > @@ -972,15 +972,15 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
> >   	 * A transient state at this point would mean some unexpected party
> >   	 * is poking at the power controls too.
> >   	 */
> > -	ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL) & mask;
> > +	ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL) & mask;
> >   	WARN_ON(ctrl != state);
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   	return enabled;
> >   }
> > -static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv)
> > +static void vlv_init_display_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -1001,13 +1001,13 @@ static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(MI_ARB_VLV, MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE);
> >   	I915_WRITE(CBR1_VLV, 0);
> > -	WARN_ON(dev_priv->rawclk_freq == 0);
> > +	WARN_ON(i915->rawclk_freq == 0);
> >   	I915_WRITE(RAWCLK_FREQ_VLV,
> > -		   DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 1000));
> > +		   DIV_ROUND_CLOSEST(i915->rawclk_freq, 1000));
> >   }
> > -static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> > +static void vlv_display_power_well_init(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> >   	enum pipe pipe;
> > @@ -1020,7 +1020,7 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> >   	 *
> >   	 * CHV DPLL B/C have some issues if VGA mode is enabled.
> >   	 */
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		u32 val = I915_READ(DPLL(pipe));
> >   		val |= DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
> > @@ -1030,71 +1030,71 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(DPLL(pipe), val);
> >   	}
> > -	vlv_init_display_clock_gating(dev_priv);
> > +	vlv_init_display_clock_gating(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	valleyview_enable_display_irqs(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	valleyview_enable_display_irqs(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/*
> >   	 * During driver initialization/resume we can avoid restoring the
> >   	 * part of the HW/SW state that will be inited anyway explicitly.
> >   	 */
> > -	if (dev_priv->power_domains.initializing)
> > +	if (i915->power_domains.initializing)
> >   		return;
> > -	intel_hpd_init(dev_priv);
> > +	intel_hpd_init(i915);
> >   	/* Re-enable the ADPA, if we have one */
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		if (encoder->type == INTEL_OUTPUT_ANALOG)
> >   			intel_crt_reset(&encoder->base);
> >   	}
> > -	i915_redisable_vga_power_on(dev_priv);
> > +	i915_redisable_vga_power_on(i915);
> > -	intel_pps_unlock_regs_wa(dev_priv);
> > +	intel_pps_unlock_regs_wa(i915);
> >   }
> > -static void vlv_display_power_well_deinit(struct drm_i915_private *dev_priv)
> > +static void vlv_display_power_well_deinit(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	valleyview_disable_display_irqs(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	valleyview_disable_display_irqs(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	/* make sure we're done processing display irqs */
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	synchronize_irq(i915->drm.irq);
> > -	intel_power_sequencer_reset(dev_priv);
> > +	intel_power_sequencer_reset(i915);
> >   	/* Prevent us from re-enabling polling on accident in late suspend */
> > -	if (!dev_priv->drm.dev->power.is_suspended)
> > -		intel_hpd_poll_init(dev_priv);
> > +	if (!i915->drm.dev->power.is_suspended)
> > +		intel_hpd_poll_init(i915);
> >   }
> > -static void vlv_display_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void vlv_display_power_well_enable(struct drm_i915_private *i915,
> >   					  struct i915_power_well *power_well)
> >   {
> > -	vlv_set_power_well(dev_priv, power_well, true);
> > +	vlv_set_power_well(i915, power_well, true);
> > -	vlv_display_power_well_init(dev_priv);
> > +	vlv_display_power_well_init(i915);
> >   }
> > -static void vlv_display_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void vlv_display_power_well_disable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> > -	vlv_display_power_well_deinit(dev_priv);
> > +	vlv_display_power_well_deinit(i915);
> > -	vlv_set_power_well(dev_priv, power_well, false);
> > +	vlv_set_power_well(i915, power_well, false);
> >   }
> > -static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> >   	/* since ref/cri clock was enabled */
> >   	udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
> > -	vlv_set_power_well(dev_priv, power_well, true);
> > +	vlv_set_power_well(i915, power_well, true);
> >   	/*
> >   	 * From VLV2A0_DP_eDP_DPIO_driver_vbios_notes_10.docx -
> > @@ -1110,31 +1110,31 @@ static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) | DPIO_CMNRST);
> >   }
> > -static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> >   					    struct i915_power_well *power_well)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe)
> > -		assert_pll_disabled(dev_priv, pipe);
> > +	for_each_pipe(i915, pipe)
> > +		assert_pll_disabled(i915, pipe);
> >   	/* Assert common reset */
> >   	I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) & ~DPIO_CMNRST);
> > -	vlv_set_power_well(dev_priv, power_well, false);
> > +	vlv_set_power_well(i915, power_well, false);
> >   }
> >   #define POWER_DOMAIN_MASK (GENMASK_ULL(POWER_DOMAIN_NUM - 1, 0))
> >   #define BITS_SET(val, bits) (((val) & (bits)) == (bits))
> > -static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> > +static void assert_chv_phy_status(struct drm_i915_private *i915)
> >   {
> >   	struct i915_power_well *cmn_bc =
> > -		lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> > +		lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> >   	struct i915_power_well *cmn_d =
> > -		lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
> > -	u32 phy_control = dev_priv->chv_phy_control;
> > +		lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D);
> > +	u32 phy_control = i915->chv_phy_control;
> >   	u32 phy_status = 0;
> >   	u32 phy_status_mask = 0xffffffff;
> > @@ -1145,7 +1145,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> >   	 * reset (ie. the power well has been disabled at
> >   	 * least once).
> >   	 */
> > -	if (!dev_priv->chv_phy_assert[DPIO_PHY0])
> > +	if (!i915->chv_phy_assert[DPIO_PHY0])
> >   		phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH0) |
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 0) |
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 1) |
> > @@ -1153,12 +1153,12 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 0) |
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1));
> > -	if (!dev_priv->chv_phy_assert[DPIO_PHY1])
> > +	if (!i915->chv_phy_assert[DPIO_PHY1])
> >   		phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY1, DPIO_CH0) |
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 0) |
> >   				     PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 1));
> > -	if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
> > +	if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) {
> >   		phy_status |= PHY_POWERGOOD(DPIO_PHY0);
> >   		/* this assumes override is only used to enable lanes */
> > @@ -1199,7 +1199,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> >   			phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1);
> >   	}
> > -	if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
> > +	if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) {
> >   		phy_status |= PHY_POWERGOOD(DPIO_PHY1);
> >   		/* this assumes override is only used to enable lanes */
> > @@ -1224,19 +1224,19 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
> >   	 * The PHY may be busy with some initial calibration and whatnot,
> >   	 * so the power state can take a while to actually change.
> >   	 */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DISPLAY_PHY_STATUS,
> >   				    phy_status_mask,
> >   				    phy_status,
> >   				    10))
> >   		DRM_ERROR("Unexpected PHY_STATUS 0x%08x, expected 0x%08x (PHY_CONTROL=0x%08x)\n",
> >   			  I915_READ(DISPLAY_PHY_STATUS) & phy_status_mask,
> > -			   phy_status, dev_priv->chv_phy_control);
> > +			   phy_status, i915->chv_phy_control);
> >   }
> >   #undef BITS_SET
> > -static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *i915,
> >   					   struct i915_power_well *power_well)
> >   {
> >   	enum dpio_phy phy;
> > @@ -1256,51 +1256,51 @@ static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
> >   	/* since ref/cri clock was enabled */
> >   	udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
> > -	vlv_set_power_well(dev_priv, power_well, true);
> > +	vlv_set_power_well(i915, power_well, true);
> >   	/* Poll for phypwrgood signal */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DISPLAY_PHY_STATUS,
> >   				    PHY_POWERGOOD(phy),
> >   				    PHY_POWERGOOD(phy),
> >   				    1))
> >   		DRM_ERROR("Display PHY %d is not power up\n", phy);
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Enable dynamic power down */
> > -	tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW28);
> > +	tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW28);
> >   	tmp |= DPIO_DYNPWRDOWNEN_CH0 | DPIO_CL1POWERDOWNEN |
> >   		DPIO_SUS_CLK_CONFIG_GATE_CLKREQ;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW28, tmp);
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW28, tmp);
> >   	if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
> > -		tmp = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW6_CH1);
> > +		tmp = vlv_dpio_read(i915, pipe, _CHV_CMN_DW6_CH1);
> >   		tmp |= DPIO_DYNPWRDOWNEN_CH1;
> > -		vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW6_CH1, tmp);
> > +		vlv_dpio_write(i915, pipe, _CHV_CMN_DW6_CH1, tmp);
> >   	} else {
> >   		/*
> >   		 * Force the non-existing CL2 off. BXT does this
> >   		 * too, so maybe it saves some power even though
> >   		 * CL2 doesn't exist?
> >   		 */
> > -		tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW30);
> > +		tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW30);
> >   		tmp |= DPIO_CL2_LDOFUSE_PWRENB;
> > -		vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW30, tmp);
> > +		vlv_dpio_write(i915, pipe, CHV_CMN_DW30, tmp);
> >   	}
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> > -	dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy);
> > -	I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> > +	i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy);
> > +	I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
> >   	DRM_DEBUG_KMS("Enabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
> > -		      phy, dev_priv->chv_phy_control);
> > +		      phy, i915->chv_phy_control);
> > -	assert_chv_phy_status(dev_priv);
> > +	assert_chv_phy_status(i915);
> >   }
> > -static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *i915,
> >   					    struct i915_power_well *power_well)
> >   {
> >   	enum dpio_phy phy;
> > @@ -1310,28 +1310,28 @@ static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
> >   	if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
> >   		phy = DPIO_PHY0;
> > -		assert_pll_disabled(dev_priv, PIPE_A);
> > -		assert_pll_disabled(dev_priv, PIPE_B);
> > +		assert_pll_disabled(i915, PIPE_A);
> > +		assert_pll_disabled(i915, PIPE_B);
> >   	} else {
> >   		phy = DPIO_PHY1;
> > -		assert_pll_disabled(dev_priv, PIPE_C);
> > +		assert_pll_disabled(i915, PIPE_C);
> >   	}
> > -	dev_priv->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy);
> > -	I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> > +	i915->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy);
> > +	I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
> > -	vlv_set_power_well(dev_priv, power_well, false);
> > +	vlv_set_power_well(i915, power_well, false);
> >   	DRM_DEBUG_KMS("Disabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
> > -		      phy, dev_priv->chv_phy_control);
> > +		      phy, i915->chv_phy_control);
> >   	/* PHY is fully reset now, so we can enable the PHY state asserts */
> > -	dev_priv->chv_phy_assert[phy] = true;
> > +	i915->chv_phy_assert[phy] = true;
> > -	assert_chv_phy_status(dev_priv);
> > +	assert_chv_phy_status(i915);
> >   }
> > -static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> > +static void assert_chv_phy_powergate(struct drm_i915_private *i915, enum dpio_phy phy,
> >   				     enum dpio_channel ch, bool override, unsigned int mask)
> >   {
> >   	enum pipe pipe = phy == DPIO_PHY0 ? PIPE_A : PIPE_C;
> > @@ -1344,7 +1344,7 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> >   	 * reset (ie. the power well has been disabled at
> >   	 * least once).
> >   	 */
> > -	if (!dev_priv->chv_phy_assert[phy])
> > +	if (!i915->chv_phy_assert[phy])
> >   		return;
> >   	if (ch == DPIO_CH0)
> > @@ -1352,9 +1352,9 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> >   	else
> >   		reg = _CHV_CMN_DW6_CH1;
> > -	vlv_dpio_get(dev_priv);
> > -	val = vlv_dpio_read(dev_priv, pipe, reg);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_get(i915);
> > +	val = vlv_dpio_read(i915, pipe, reg);
> > +	vlv_dpio_put(i915);
> >   	/*
> >   	 * This assumes !override is only used when the port is disabled.
> > @@ -1393,30 +1393,30 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi
> >   	     reg, val);
> >   }
> > -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> > +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy,
> >   			  enum dpio_channel ch, bool override)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	bool was_override;
> >   	mutex_lock(&power_domains->lock);
> > -	was_override = dev_priv->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > +	was_override = i915->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> >   	if (override == was_override)
> >   		goto out;
> >   	if (override)
> > -		dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > +		i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> >   	else
> > -		dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > +		i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > -	I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> > +	I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
> >   	DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d (DPIO_PHY_CONTROL=0x%08x)\n",
> > -		      phy, ch, dev_priv->chv_phy_control);
> > +		      phy, ch, i915->chv_phy_control);
> > -	assert_chv_phy_status(dev_priv);
> > +	assert_chv_phy_status(i915);
> >   out:
> >   	mutex_unlock(&power_domains->lock);
> > @@ -1427,43 +1427,43 @@ bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> >   void chv_phy_powergate_lanes(struct intel_encoder *encoder,
> >   			     bool override, unsigned int mask)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	enum dpio_phy phy = vlv_dport_to_phy(enc_to_dig_port(&encoder->base));
> >   	enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base));
> >   	mutex_lock(&power_domains->lock);
> > -	dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch);
> > -	dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch);
> > +	i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch);
> > +	i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch);
> >   	if (override)
> > -		dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > +		i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> >   	else
> > -		dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > +		i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
> > -	I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> > +	I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
> >   	DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d lanes 0x%x (PHY_CONTROL=0x%08x)\n",
> > -		      phy, ch, mask, dev_priv->chv_phy_control);
> > +		      phy, ch, mask, i915->chv_phy_control);
> > -	assert_chv_phy_status(dev_priv);
> > +	assert_chv_phy_status(i915);
> > -	assert_chv_phy_powergate(dev_priv, phy, ch, override, mask);
> > +	assert_chv_phy_powergate(i915, phy, ch, override, mask);
> >   	mutex_unlock(&power_domains->lock);
> >   }
> > -static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv,
> > +static bool chv_pipe_power_well_enabled(struct drm_i915_private *i915,
> >   					struct i915_power_well *power_well)
> >   {
> >   	enum pipe pipe = PIPE_A;
> >   	bool enabled;
> >   	u32 state, ctrl;
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> > -	state = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe);
> > +	state = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe);
> >   	/*
> >   	 * We only ever set the power-on and power-gate states, anything
> >   	 * else is unexpected.
> > @@ -1475,15 +1475,15 @@ static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv,
> >   	 * A transient state at this point would mean some unexpected party
> >   	 * is poking at the power controls too.
> >   	 */
> > -	ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe);
> > +	ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe);
> >   	WARN_ON(ctrl << 16 != state);
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   	return enabled;
> >   }
> > -static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv,
> > +static void chv_set_pipe_power_well(struct drm_i915_private *i915,
> >   				    struct i915_power_well *power_well,
> >   				    bool enable)
> >   {
> > @@ -1493,44 +1493,44 @@ static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv,
> >   	state = enable ? DP_SSS_PWR_ON(pipe) : DP_SSS_PWR_GATE(pipe);
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> >   #define COND \
> > -	((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state)
> > +	((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state)
> >   	if (COND)
> >   		goto out;
> > -	ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +	ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> >   	ctrl &= ~DP_SSC_MASK(pipe);
> >   	ctrl |= enable ? DP_SSC_PWR_ON(pipe) : DP_SSC_PWR_GATE(pipe);
> > -	vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, ctrl);
> > +	vlv_punit_write(i915, PUNIT_REG_DSPSSPM, ctrl);
> >   	if (wait_for(COND, 100))
> >   		DRM_ERROR("timeout setting power well state %08x (%08x)\n",
> >   			  state,
> > -			  vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM));
> > +			  vlv_punit_read(i915, PUNIT_REG_DSPSSPM));
> >   #undef COND
> >   out:
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   }
> > -static void chv_pipe_power_well_enable(struct drm_i915_private *dev_priv,
> > +static void chv_pipe_power_well_enable(struct drm_i915_private *i915,
> >   				       struct i915_power_well *power_well)
> >   {
> > -	chv_set_pipe_power_well(dev_priv, power_well, true);
> > +	chv_set_pipe_power_well(i915, power_well, true);
> > -	vlv_display_power_well_init(dev_priv);
> > +	vlv_display_power_well_init(i915);
> >   }
> > -static void chv_pipe_power_well_disable(struct drm_i915_private *dev_priv,
> > +static void chv_pipe_power_well_disable(struct drm_i915_private *i915,
> >   					struct i915_power_well *power_well)
> >   {
> > -	vlv_display_power_well_deinit(dev_priv);
> > +	vlv_display_power_well_deinit(i915);
> > -	chv_set_pipe_power_well(dev_priv, power_well, false);
> > +	chv_set_pipe_power_well(i915, power_well, false);
> >   }
> >   static u64 __async_put_domains_mask(struct i915_power_domains *power_domains)
> > @@ -1627,10 +1627,10 @@ async_put_domains_clear_domain(struct i915_power_domains *power_domains,
> >   }
> >   static bool
> > -intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> > +intel_display_power_grab_async_put_ref(struct drm_i915_private *i915,
> >   				       enum intel_display_power_domain domain)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	bool ret = false;
> >   	if (!(async_put_domains_mask(power_domains) & BIT_ULL(domain)))
> > @@ -1644,7 +1644,7 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> >   		goto out_verify;
> >   	cancel_delayed_work(&power_domains->async_put_work);
> > -	intel_runtime_pm_put_raw(dev_priv,
> > +	intel_runtime_pm_put_raw(i915,
> >   				 fetch_and_zero(&power_domains->async_put_wakeref));
> >   out_verify:
> >   	verify_async_put_domains_state(power_domains);
> > @@ -1653,24 +1653,24 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -__intel_display_power_get_domain(struct drm_i915_private *dev_priv,
> > +__intel_display_power_get_domain(struct drm_i915_private *i915,
> >   				 enum intel_display_power_domain domain)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *power_well;
> > -	if (intel_display_power_grab_async_put_ref(dev_priv, domain))
> > +	if (intel_display_power_grab_async_put_ref(i915, domain))
> >   		return;
> > -	for_each_power_domain_well(dev_priv, power_well, BIT_ULL(domain))
> > -		intel_power_well_get(dev_priv, power_well);
> > +	for_each_power_domain_well(i915, power_well, BIT_ULL(domain))
> > +		intel_power_well_get(i915, power_well);
> >   	power_domains->domain_use_count[domain]++;
> >   }
> >   /**
> >    * intel_display_power_get - grab a power domain reference
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to reference
> >    *
> >    * This function grabs a power domain reference for @domain and ensures that the
> > @@ -1680,14 +1680,14 @@ __intel_display_power_get_domain(struct drm_i915_private *dev_priv,
> >    * Any power domain reference obtained by this function must have a symmetric
> >    * call to intel_display_power_put() to release the reference again.
> >    */
> > -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> > +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915,
> >   					enum intel_display_power_domain domain)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > -	intel_wakeref_t wakeref = intel_runtime_pm_get(dev_priv);
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> > +	intel_wakeref_t wakeref = intel_runtime_pm_get(i915);
> >   	mutex_lock(&power_domains->lock);
> > -	__intel_display_power_get_domain(dev_priv, domain);
> > +	__intel_display_power_get_domain(i915, domain);
> >   	mutex_unlock(&power_domains->lock);
> >   	return wakeref;
> > @@ -1695,7 +1695,7 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_display_power_get_if_enabled - grab a reference for an enabled display power domain
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to reference
> >    *
> >    * This function grabs a power domain reference for @domain and ensures that the
> > @@ -1706,21 +1706,21 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> >    * call to intel_display_power_put() to release the reference again.
> >    */
> >   intel_wakeref_t
> > -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> > +intel_display_power_get_if_enabled(struct drm_i915_private *i915,
> >   				   enum intel_display_power_domain domain)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	intel_wakeref_t wakeref;
> >   	bool is_enabled;
> > -	wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
> > +	wakeref = intel_runtime_pm_get_if_in_use(i915);
> >   	if (!wakeref)
> >   		return false;
> >   	mutex_lock(&power_domains->lock);
> > -	if (__intel_display_power_is_enabled(dev_priv, domain)) {
> > -		__intel_display_power_get_domain(dev_priv, domain);
> > +	if (__intel_display_power_is_enabled(i915, domain)) {
> > +		__intel_display_power_get_domain(i915, domain);
> >   		is_enabled = true;
> >   	} else {
> >   		is_enabled = false;
> > @@ -1729,7 +1729,7 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> >   	mutex_unlock(&power_domains->lock);
> >   	if (!is_enabled) {
> > -		intel_runtime_pm_put(dev_priv, wakeref);
> > +		intel_runtime_pm_put(i915, wakeref);
> >   		wakeref = 0;
> >   	}
> > @@ -1737,14 +1737,14 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -__intel_display_power_put_domain(struct drm_i915_private *dev_priv,
> > +__intel_display_power_put_domain(struct drm_i915_private *i915,
> >   				 enum intel_display_power_domain domain)
> >   {
> >   	struct i915_power_domains *power_domains;
> >   	struct i915_power_well *power_well;
> >   	const char *name = intel_display_power_domain_str(domain);
> > -	power_domains = &dev_priv->power_domains;
> > +	power_domains = &i915->power_domains;
> >   	WARN(!power_domains->domain_use_count[domain],
> >   	     "Use count on domain %s is already zero\n",
> > @@ -1755,23 +1755,23 @@ __intel_display_power_put_domain(struct drm_i915_private *dev_priv,
> >   	power_domains->domain_use_count[domain]--;
> > -	for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain))
> > -		intel_power_well_put(dev_priv, power_well);
> > +	for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain))
> > +		intel_power_well_put(i915, power_well);
> >   }
> > -static void __intel_display_power_put(struct drm_i915_private *dev_priv,
> > +static void __intel_display_power_put(struct drm_i915_private *i915,
> >   				      enum intel_display_power_domain domain)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	mutex_lock(&power_domains->lock);
> > -	__intel_display_power_put_domain(dev_priv, domain);
> > +	__intel_display_power_put_domain(i915, domain);
> >   	mutex_unlock(&power_domains->lock);
> >   }
> >   /**
> >    * intel_display_power_put_unchecked - release an unchecked power domain reference
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to reference
> >    *
> >    * This function drops the power domain reference obtained by
> > @@ -1782,11 +1782,11 @@ static void __intel_display_power_put(struct drm_i915_private *dev_priv,
> >    * new code, as the correctness of its use cannot be checked. Always use
> >    * intel_display_power_put() instead.
> >    */
> > -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv,
> > +void intel_display_power_put_unchecked(struct drm_i915_private *i915,
> >   				       enum intel_display_power_domain domain)
> >   {
> > -	__intel_display_power_put(dev_priv, domain);
> > -	intel_runtime_pm_put_unchecked(dev_priv);
> > +	__intel_display_power_put(i915, domain);
> > +	intel_runtime_pm_put_unchecked(i915);
> >   }
> >   static void
> > @@ -1803,7 +1803,7 @@ queue_async_put_domains_work(struct i915_power_domains *power_domains,
> >   static void
> >   release_async_put_domains(struct i915_power_domains *power_domains, u64 mask)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(power_domains, struct drm_i915_private,
> >   			     power_domains);
> >   	enum intel_display_power_domain domain;
> > @@ -1814,26 +1814,26 @@ release_async_put_domains(struct i915_power_domains *power_domains, u64 mask)
> >   	 * wakeref to make the state checker happy about the HW access during
> >   	 * power well disabling.
> >   	 */
> > -	assert_rpm_raw_wakeref_held(dev_priv);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	assert_rpm_raw_wakeref_held(i915);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	for_each_power_domain(domain, mask) {
> >   		/* Clear before put, so put's sanity check is happy. */
> >   		async_put_domains_clear_domain(power_domains, domain);
> > -		__intel_display_power_put_domain(dev_priv, domain);
> > +		__intel_display_power_put_domain(i915, domain);
> >   	}
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   }
> >   static void
> >   intel_display_power_put_async_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private,
> >   			     power_domains.async_put_work.work);
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > -	intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(dev_priv);
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> > +	intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(i915);
> >   	intel_wakeref_t old_work_wakeref = 0;
> >   	mutex_lock(&power_domains->lock);
> > @@ -1863,9 +1863,9 @@ intel_display_power_put_async_work(struct work_struct *work)
> >   	mutex_unlock(&power_domains->lock);
> >   	if (old_work_wakeref)
> > -		intel_runtime_pm_put_raw(dev_priv, old_work_wakeref);
> > +		intel_runtime_pm_put_raw(i915, old_work_wakeref);
> >   	if (new_work_wakeref)
> > -		intel_runtime_pm_put_raw(dev_priv, new_work_wakeref);
> > +		intel_runtime_pm_put_raw(i915, new_work_wakeref);
> >   }
> >   /**
> > @@ -1974,7 +1974,7 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915)
> >   #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
> >   /**
> >    * intel_display_power_put - release a power domain reference
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @domain: power domain to reference
> >    * @wakeref: wakeref acquired for the reference that is being released
> >    *
> > @@ -1982,12 +1982,12 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915)
> >    * intel_display_power_get() and might power down the corresponding hardware
> >    * block right away if this is the last reference.
> >    */
> > -void intel_display_power_put(struct drm_i915_private *dev_priv,
> > +void intel_display_power_put(struct drm_i915_private *i915,
> >   			     enum intel_display_power_domain domain,
> >   			     intel_wakeref_t wakeref)
> >   {
> > -	__intel_display_power_put(dev_priv, domain);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	__intel_display_power_put(i915, domain);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   }
> >   #endif
> > @@ -2655,14 +2655,14 @@ static const struct i915_power_well_desc chv_power_wells[] = {
> >   	},
> >   };
> > -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915,
> >   					 enum i915_power_well_id power_well_id)
> >   {
> >   	struct i915_power_well *power_well;
> >   	bool ret;
> > -	power_well = lookup_power_well(dev_priv, power_well_id);
> > -	ret = power_well->desc->ops->is_enabled(dev_priv, power_well);
> > +	power_well = lookup_power_well(i915, power_well_id);
> > +	ret = power_well->desc->ops->is_enabled(i915, power_well);
> >   	return ret;
> >   }
> > @@ -3352,7 +3352,7 @@ static const struct i915_power_well_desc icl_power_wells[] = {
> >   };
> >   static int
> > -sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv,
> > +sanitize_disable_power_well_option(const struct drm_i915_private *i915,
> >   				   int disable_power_well)
> >   {
> >   	if (disable_power_well >= 0)
> > @@ -3361,14 +3361,14 @@ sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv,
> >   	return 1;
> >   }
> > -static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv,
> > +static u32 get_allowed_dc_mask(const struct drm_i915_private *i915,
> >   			       int enable_dc)
> >   {
> >   	u32 mask;
> >   	int requested_dc;
> >   	int max_dc;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		max_dc = 2;
> >   		/*
> >   		 * DC9 has a separate HW flow from the rest of the DC states,
> > @@ -3376,10 +3376,10 @@ static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv,
> >   		 * suspend/resume, so allow it unconditionally.
> >   		 */
> >   		mask = DC_STATE_EN_DC9;
> > -	} else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv)) {
> > +	} else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915)) {
> >   		max_dc = 2;
> >   		mask = 0;
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > +	} else if (IS_GEN9_LP(i915)) {
> >   		max_dc = 1;
> >   		mask = DC_STATE_EN_DC9;
> >   	} else {
> > @@ -3451,21 +3451,21 @@ __set_power_wells(struct i915_power_domains *power_domains,
> >   /**
> >    * intel_power_domains_init - initializes the power domain structures
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> > - * Initializes the power domain structures for @dev_priv depending upon the
> > + * Initializes the power domain structures for @i915 depending upon the
> >    * supported platform.
> >    */
> > -int intel_power_domains_init(struct drm_i915_private *dev_priv)
> > +int intel_power_domains_init(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	int err;
> >   	i915_modparams.disable_power_well =
> > -		sanitize_disable_power_well_option(dev_priv,
> > +		sanitize_disable_power_well_option(i915,
> >   						   i915_modparams.disable_power_well);
> > -	dev_priv->csr.allowed_dc_mask =
> > -		get_allowed_dc_mask(dev_priv, i915_modparams.enable_dc);
> > +	i915->csr.allowed_dc_mask =
> > +		get_allowed_dc_mask(i915, i915_modparams.enable_dc);
> >   	BUILD_BUG_ON(POWER_DOMAIN_NUM > 64);
> > @@ -3478,9 +3478,9 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
> >   	 * The enabling order will be from lower to higher indexed wells,
> >   	 * the disabling order is reversed.
> >   	 */
> > -	if (IS_GEN(dev_priv, 11)) {
> > +	if (IS_GEN(i915, 11)) {
> >   		err = set_power_wells(power_domains, icl_power_wells);
> > -	} else if (IS_CANNONLAKE(dev_priv)) {
> > +	} else if (IS_CANNONLAKE(i915)) {
> >   		err = set_power_wells(power_domains, cnl_power_wells);
> >   		/*
> > @@ -3489,23 +3489,23 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
> >   		 * timeouts, lets remove them from the list
> >   		 * for the SKUs without port F.
> >   		 */
> > -		if (!IS_CNL_WITH_PORT_F(dev_priv))
> > +		if (!IS_CNL_WITH_PORT_F(i915))
> >   			power_domains->power_well_count -= 2;
> > -	} else if (IS_GEMINILAKE(dev_priv)) {
> > +	} else if (IS_GEMINILAKE(i915)) {
> >   		err = set_power_wells(power_domains, glk_power_wells);
> > -	} else if (IS_BROXTON(dev_priv)) {
> > +	} else if (IS_BROXTON(i915)) {
> >   		err = set_power_wells(power_domains, bxt_power_wells);
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > +	} else if (IS_GEN9_BC(i915)) {
> >   		err = set_power_wells(power_domains, skl_power_wells);
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_CHERRYVIEW(i915)) {
> >   		err = set_power_wells(power_domains, chv_power_wells);
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > +	} else if (IS_BROADWELL(i915)) {
> >   		err = set_power_wells(power_domains, bdw_power_wells);
> > -	} else if (IS_HASWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915)) {
> >   		err = set_power_wells(power_domains, hsw_power_wells);
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		err = set_power_wells(power_domains, vlv_power_wells);
> > -	} else if (IS_I830(dev_priv)) {
> > +	} else if (IS_I830(i915)) {
> >   		err = set_power_wells(power_domains, i830_power_wells);
> >   	} else {
> >   		err = set_power_wells(power_domains, i9xx_always_on_power_well);
> > @@ -3516,31 +3516,31 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_power_domains_cleanup - clean up power domains resources
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Release any resources acquired by intel_power_domains_init()
> >    */
> > -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv)
> > +void intel_power_domains_cleanup(struct drm_i915_private *i915)
> >   {
> > -	kfree(dev_priv->power_domains.power_wells);
> > +	kfree(i915->power_domains.power_wells);
> >   }
> > -static void intel_power_domains_sync_hw(struct drm_i915_private *dev_priv)
> > +static void intel_power_domains_sync_hw(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *power_well;
> >   	mutex_lock(&power_domains->lock);
> > -	for_each_power_well(dev_priv, power_well) {
> > -		power_well->desc->ops->sync_hw(dev_priv, power_well);
> > +	for_each_power_well(i915, power_well) {
> > +		power_well->desc->ops->sync_hw(i915, power_well);
> >   		power_well->hw_enabled =
> > -			power_well->desc->ops->is_enabled(dev_priv, power_well);
> > +			power_well->desc->ops->is_enabled(i915, power_well);
> >   	}
> >   	mutex_unlock(&power_domains->lock);
> >   }
> >   static inline
> > -bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv,
> > +bool intel_dbuf_slice_set(struct drm_i915_private *i915,
> >   			  i915_reg_t reg, bool enable)
> >   {
> >   	u32 val, status;
> > @@ -3560,30 +3560,30 @@ bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static void gen9_dbuf_enable(struct drm_i915_private *dev_priv)
> > +static void gen9_dbuf_enable(struct drm_i915_private *i915)
> >   {
> > -	intel_dbuf_slice_set(dev_priv, DBUF_CTL, true);
> > +	intel_dbuf_slice_set(i915, DBUF_CTL, true);
> >   }
> > -static void gen9_dbuf_disable(struct drm_i915_private *dev_priv)
> > +static void gen9_dbuf_disable(struct drm_i915_private *i915)
> >   {
> > -	intel_dbuf_slice_set(dev_priv, DBUF_CTL, false);
> > +	intel_dbuf_slice_set(i915, DBUF_CTL, false);
> >   }
> > -static u8 intel_dbuf_max_slices(struct drm_i915_private *dev_priv)
> > +static u8 intel_dbuf_max_slices(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return 1;
> >   	return 2;
> >   }
> > -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> > +void icl_dbuf_slices_update(struct drm_i915_private *i915,
> >   			    u8 req_slices)
> >   {
> > -	const u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices;
> > +	const u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices;
> >   	bool ret;
> > -	if (req_slices > intel_dbuf_max_slices(dev_priv)) {
> > +	if (req_slices > intel_dbuf_max_slices(i915)) {
> >   		DRM_ERROR("Invalid number of dbuf slices requested\n");
> >   		return;
> >   	}
> > @@ -3592,15 +3592,15 @@ void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> >   		return;
> >   	if (req_slices > hw_enabled_slices)
> > -		ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, true);
> > +		ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, true);
> >   	else
> > -		ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, false);
> > +		ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, false);
> >   	if (ret)
> > -		dev_priv->wm.skl_hw.ddb.enabled_slices = req_slices;
> > +		i915->wm.skl_hw.ddb.enabled_slices = req_slices;
> >   }
> > -static void icl_dbuf_enable(struct drm_i915_private *dev_priv)
> > +static void icl_dbuf_enable(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) | DBUF_POWER_REQUEST);
> >   	I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) | DBUF_POWER_REQUEST);
> > @@ -3616,10 +3616,10 @@ static void icl_dbuf_enable(struct drm_i915_private *dev_priv)
> >   		 * FIXME: for now pretend that we only have 1 slice, see
> >   		 * intel_enabled_dbuf_slices_num().
> >   		 */
> > -		dev_priv->wm.skl_hw.ddb.enabled_slices = 1;
> > +		i915->wm.skl_hw.ddb.enabled_slices = 1;
> >   }
> > -static void icl_dbuf_disable(struct drm_i915_private *dev_priv)
> > +static void icl_dbuf_disable(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) & ~DBUF_POWER_REQUEST);
> >   	I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) & ~DBUF_POWER_REQUEST);
> > @@ -3635,10 +3635,10 @@ static void icl_dbuf_disable(struct drm_i915_private *dev_priv)
> >   		 * FIXME: for now pretend that the first slice is always
> >   		 * enabled, see intel_enabled_dbuf_slices_num().
> >   		 */
> > -		dev_priv->wm.skl_hw.ddb.enabled_slices = 1;
> > +		i915->wm.skl_hw.ddb.enabled_slices = 1;
> >   }
> > -static void icl_mbus_init(struct drm_i915_private *dev_priv)
> > +static void icl_mbus_init(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -3650,7 +3650,7 @@ static void icl_mbus_init(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(MBUS_ABOX_CTL, val);
> >   }
> > -static void hsw_assert_cdclk(struct drm_i915_private *dev_priv)
> > +static void hsw_assert_cdclk(struct drm_i915_private *i915)
> >   {
> >   	u32 val = I915_READ(LCPLL_CTL);
> > @@ -3670,9 +3670,9 @@ static void hsw_assert_cdclk(struct drm_i915_private *dev_priv)
> >   		DRM_ERROR("LCPLL not using non-SSC reference\n");
> >   }
> > -static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> > +static void assert_can_disable_lcpll(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_crtc *crtc;
> >   	for_each_intel_crtc(dev, crtc)
> > @@ -3691,7 +3691,7 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> >   			"Panel power on\n");
> >   	I915_STATE_WARN(I915_READ(BLC_PWM_CPU_CTL2) & BLM_PWM_ENABLE,
> >   			"CPU PWM1 enabled\n");
> > -	if (IS_HASWELL(dev_priv))
> > +	if (IS_HASWELL(i915))
> >   		I915_STATE_WARN(I915_READ(HSW_BLC_PWM2_CTL) & BLM_PWM_ENABLE,
> >   				"CPU PWM2 enabled\n");
> >   	I915_STATE_WARN(I915_READ(BLC_PWM_PCH_CTL1) & BLM_PCH_PWM_ENABLE,
> > @@ -3707,21 +3707,21 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv)
> >   	 * gen-specific and since we only disable LCPLL after we fully disable
> >   	 * the interrupts, the check below should be enough.
> >   	 */
> > -	I915_STATE_WARN(intel_irqs_enabled(dev_priv), "IRQs enabled\n");
> > +	I915_STATE_WARN(intel_irqs_enabled(i915), "IRQs enabled\n");
> >   }
> > -static u32 hsw_read_dcomp(struct drm_i915_private *dev_priv)
> > +static u32 hsw_read_dcomp(struct drm_i915_private *i915)
> >   {
> > -	if (IS_HASWELL(dev_priv))
> > +	if (IS_HASWELL(i915))
> >   		return I915_READ(D_COMP_HSW);
> >   	else
> >   		return I915_READ(D_COMP_BDW);
> >   }
> > -static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val)
> > +static void hsw_write_dcomp(struct drm_i915_private *i915, u32 val)
> >   {
> > -	if (IS_HASWELL(dev_priv)) {
> > -		if (sandybridge_pcode_write(dev_priv,
> > +	if (IS_HASWELL(i915)) {
> > +		if (sandybridge_pcode_write(i915,
> >   					    GEN6_PCODE_WRITE_D_COMP, val))
> >   			DRM_DEBUG_KMS("Failed to write to D_COMP\n");
> >   	} else {
> > @@ -3738,12 +3738,12 @@ static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val)
> >    * register. Callers should take care of disabling all the display engine
> >    * functions, doing the mode unset, fixing interrupts, etc.
> >    */
> > -static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> > +static void hsw_disable_lcpll(struct drm_i915_private *i915,
> >   			      bool switch_to_fclk, bool allow_power_down)
> >   {
> >   	u32 val;
> > -	assert_can_disable_lcpll(dev_priv);
> > +	assert_can_disable_lcpll(i915);
> >   	val = I915_READ(LCPLL_CTL);
> > @@ -3762,16 +3762,16 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(LCPLL_CTL, val);
> >   	POSTING_READ(LCPLL_CTL);
> > -	if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL,
> > +	if (intel_wait_for_register(&i915->uncore, LCPLL_CTL,
> >   				    LCPLL_PLL_LOCK, 0, 1))
> >   		DRM_ERROR("LCPLL still locked\n");
> > -	val = hsw_read_dcomp(dev_priv);
> > +	val = hsw_read_dcomp(i915);
> >   	val |= D_COMP_COMP_DISABLE;
> > -	hsw_write_dcomp(dev_priv, val);
> > +	hsw_write_dcomp(i915, val);
> >   	ndelay(100);
> > -	if (wait_for((hsw_read_dcomp(dev_priv) &
> > +	if (wait_for((hsw_read_dcomp(i915) &
> >   		      D_COMP_RCOMP_IN_PROGRESS) == 0, 1))
> >   		DRM_ERROR("D_COMP RCOMP still in progress\n");
> > @@ -3787,7 +3787,7 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv,
> >    * Fully restores LCPLL, disallowing power down and switching back to LCPLL
> >    * source.
> >    */
> > -static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> > +static void hsw_restore_lcpll(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > @@ -3801,7 +3801,7 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> >   	 * Make sure we're not on PC8 state before disabling PC8, otherwise
> >   	 * we'll hang the machine. To prevent PC8 state, just enable force_wake.
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	if (val & LCPLL_POWER_DOWN_ALLOW) {
> >   		val &= ~LCPLL_POWER_DOWN_ALLOW;
> > @@ -3809,16 +3809,16 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> >   		POSTING_READ(LCPLL_CTL);
> >   	}
> > -	val = hsw_read_dcomp(dev_priv);
> > +	val = hsw_read_dcomp(i915);
> >   	val |= D_COMP_COMP_FORCE;
> >   	val &= ~D_COMP_COMP_DISABLE;
> > -	hsw_write_dcomp(dev_priv, val);
> > +	hsw_write_dcomp(i915, val);
> >   	val = I915_READ(LCPLL_CTL);
> >   	val &= ~LCPLL_PLL_DISABLE;
> >   	I915_WRITE(LCPLL_CTL, val);
> > -	if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL,
> > +	if (intel_wait_for_register(&i915->uncore, LCPLL_CTL,
> >   				    LCPLL_PLL_LOCK, LCPLL_PLL_LOCK, 5))
> >   		DRM_ERROR("LCPLL not locked yet\n");
> > @@ -3832,10 +3832,10 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> >   			DRM_ERROR("Switching back to LCPLL failed\n");
> >   	}
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> > -	intel_update_cdclk(dev_priv);
> > -	intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK");
> > +	intel_update_cdclk(i915);
> > +	intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK");
> >   }
> >   /*
> > @@ -3861,45 +3861,45 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv)
> >    * For more, read "Display Sequences for Package C8" on the hardware
> >    * documentation.
> >    */
> > -void hsw_enable_pc8(struct drm_i915_private *dev_priv)
> > +void hsw_enable_pc8(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> >   	DRM_DEBUG_KMS("Enabling package C8+\n");
> > -	if (HAS_PCH_LPT_LP(dev_priv)) {
> > +	if (HAS_PCH_LPT_LP(i915)) {
> >   		val = I915_READ(SOUTH_DSPCLK_GATE_D);
> >   		val &= ~PCH_LP_PARTITION_LEVEL_DISABLE;
> >   		I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> >   	}
> > -	lpt_disable_clkout_dp(dev_priv);
> > -	hsw_disable_lcpll(dev_priv, true, true);
> > +	lpt_disable_clkout_dp(i915);
> > +	hsw_disable_lcpll(i915, true, true);
> >   }
> > -void hsw_disable_pc8(struct drm_i915_private *dev_priv)
> > +void hsw_disable_pc8(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> >   	DRM_DEBUG_KMS("Disabling package C8+\n");
> > -	hsw_restore_lcpll(dev_priv);
> > -	intel_init_pch_refclk(dev_priv);
> > +	hsw_restore_lcpll(i915);
> > +	intel_init_pch_refclk(i915);
> > -	if (HAS_PCH_LPT_LP(dev_priv)) {
> > +	if (HAS_PCH_LPT_LP(i915)) {
> >   		val = I915_READ(SOUTH_DSPCLK_GATE_D);
> >   		val |= PCH_LP_PARTITION_LEVEL_DISABLE;
> >   		I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> >   	}
> >   }
> > -static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv,
> > +static void intel_pch_reset_handshake(struct drm_i915_private *i915,
> >   				      bool enable)
> >   {
> >   	i915_reg_t reg;
> >   	u32 reset_bits, val;
> > -	if (IS_IVYBRIDGE(dev_priv)) {
> > +	if (IS_IVYBRIDGE(i915)) {
> >   		reg = GEN7_MSG_CTL;
> >   		reset_bits = WAIT_FOR_PCH_FLR_ACK | WAIT_FOR_PCH_RESET_ACK;
> >   	} else {
> > @@ -3917,46 +3917,46 @@ static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(reg, val);
> >   }
> > -static void skl_display_core_init(struct drm_i915_private *dev_priv,
> > +static void skl_display_core_init(struct drm_i915_private *i915,
> >   				  bool resume)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/* enable PCH reset handshake */
> > -	intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> > +	intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
> >   	/* enable PG1 and Misc I/O */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_enable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_enable(i915, well);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_MISC_IO);
> > -	intel_power_well_enable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_MISC_IO);
> > +	intel_power_well_enable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> > -	intel_cdclk_init(dev_priv);
> > +	intel_cdclk_init(i915);
> > -	gen9_dbuf_enable(dev_priv);
> > +	gen9_dbuf_enable(i915);
> > -	if (resume && dev_priv->csr.dmc_payload)
> > -		intel_csr_load_program(dev_priv);
> > +	if (resume && i915->csr.dmc_payload)
> > +		intel_csr_load_program(i915);
> >   }
> > -static void skl_display_core_uninit(struct drm_i915_private *dev_priv)
> > +static void skl_display_core_uninit(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> > -	gen9_dbuf_disable(dev_priv);
> > +	gen9_dbuf_disable(i915);
> > -	intel_cdclk_uninit(dev_priv);
> > +	intel_cdclk_uninit(i915);
> >   	/* The spec doesn't call for removing the reset handshake flag */
> >   	/* disable PG1 and Misc I/O */
> > @@ -3969,21 +3969,21 @@ static void skl_display_core_uninit(struct drm_i915_private *dev_priv)
> >   	 * Note that even though the driver's request is removed power well 1
> >   	 * may stay enabled after this due to DMC's own request on it.
> >   	 */
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_disable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_disable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	usleep_range(10, 30);		/* 10 us delay per Bspec */
> >   }
> > -void bxt_display_core_init(struct drm_i915_private *dev_priv,
> > +void bxt_display_core_init(struct drm_i915_private *i915,
> >   			   bool resume)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/*
> >   	 * NDE_RSTWRN_OPT RST PCH Handshake En must always be 0b on BXT
> > @@ -3991,34 +3991,34 @@ void bxt_display_core_init(struct drm_i915_private *dev_priv,
> >   	 * Move the handshake programming to initialization sequence.
> >   	 * Previously was left up to BIOS.
> >   	 */
> > -	intel_pch_reset_handshake(dev_priv, false);
> > +	intel_pch_reset_handshake(i915, false);
> >   	/* Enable PG1 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_enable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_enable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> > -	intel_cdclk_init(dev_priv);
> > +	intel_cdclk_init(i915);
> > -	gen9_dbuf_enable(dev_priv);
> > +	gen9_dbuf_enable(i915);
> > -	if (resume && dev_priv->csr.dmc_payload)
> > -		intel_csr_load_program(dev_priv);
> > +	if (resume && i915->csr.dmc_payload)
> > +		intel_csr_load_program(i915);
> >   }
> > -void bxt_display_core_uninit(struct drm_i915_private *dev_priv)
> > +void bxt_display_core_uninit(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> > -	gen9_dbuf_disable(dev_priv);
> > +	gen9_dbuf_disable(i915);
> > -	intel_cdclk_uninit(dev_priv);
> > +	intel_cdclk_uninit(i915);
> >   	/* The spec doesn't call for removing the reset handshake flag */
> > @@ -4029,60 +4029,60 @@ void bxt_display_core_uninit(struct drm_i915_private *dev_priv)
> >   	 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_disable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_disable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	usleep_range(10, 30);		/* 10 us delay per Bspec */
> >   }
> > -static void cnl_display_core_init(struct drm_i915_private *dev_priv, bool resume)
> > +static void cnl_display_core_init(struct drm_i915_private *i915, bool resume)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/* 1. Enable PCH Reset Handshake */
> > -	intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> > +	intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
> >   	/* 2-3. */
> > -	intel_combo_phy_init(dev_priv);
> > +	intel_combo_phy_init(i915);
> >   	/*
> >   	 * 4. Enable Power Well 1 (PG1).
> >   	 *    The AUX IO power wells will be enabled on demand.
> >   	 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_enable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_enable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	/* 5. Enable CD clock */
> > -	intel_cdclk_init(dev_priv);
> > +	intel_cdclk_init(i915);
> >   	/* 6. Enable DBUF */
> > -	gen9_dbuf_enable(dev_priv);
> > +	gen9_dbuf_enable(i915);
> > -	if (resume && dev_priv->csr.dmc_payload)
> > -		intel_csr_load_program(dev_priv);
> > +	if (resume && i915->csr.dmc_payload)
> > +		intel_csr_load_program(i915);
> >   }
> > -static void cnl_display_core_uninit(struct drm_i915_private *dev_priv)
> > +static void cnl_display_core_uninit(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/* 1. Disable all display engine functions -> aready done */
> >   	/* 2. Disable DBUF */
> > -	gen9_dbuf_disable(dev_priv);
> > +	gen9_dbuf_disable(i915);
> >   	/* 3. Disable CD clock */
> > -	intel_cdclk_uninit(dev_priv);
> > +	intel_cdclk_uninit(i915);
> >   	/*
> >   	 * 4. Disable Power Well 1 (PG1).
> > @@ -4090,66 +4090,66 @@ static void cnl_display_core_uninit(struct drm_i915_private *dev_priv)
> >   	 *    disabled at this point.
> >   	 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_disable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_disable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	usleep_range(10, 30);		/* 10 us delay per Bspec */
> >   	/* 5. */
> > -	intel_combo_phy_uninit(dev_priv);
> > +	intel_combo_phy_uninit(i915);
> >   }
> > -void icl_display_core_init(struct drm_i915_private *dev_priv,
> > +void icl_display_core_init(struct drm_i915_private *i915,
> >   			   bool resume)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/* 1. Enable PCH reset handshake. */
> > -	intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
> > +	intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
> >   	/* 2. Initialize all combo phys */
> > -	intel_combo_phy_init(dev_priv);
> > +	intel_combo_phy_init(i915);
> >   	/*
> >   	 * 3. Enable Power Well 1 (PG1).
> >   	 *    The AUX IO power wells will be enabled on demand.
> >   	 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_enable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_enable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	/* 4. Enable CDCLK. */
> > -	intel_cdclk_init(dev_priv);
> > +	intel_cdclk_init(i915);
> >   	/* 5. Enable DBUF. */
> > -	icl_dbuf_enable(dev_priv);
> > +	icl_dbuf_enable(i915);
> >   	/* 6. Setup MBUS. */
> > -	icl_mbus_init(dev_priv);
> > +	icl_mbus_init(i915);
> > -	if (resume && dev_priv->csr.dmc_payload)
> > -		intel_csr_load_program(dev_priv);
> > +	if (resume && i915->csr.dmc_payload)
> > +		intel_csr_load_program(i915);
> >   }
> > -void icl_display_core_uninit(struct drm_i915_private *dev_priv)
> > +void icl_display_core_uninit(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *well;
> > -	gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
> > +	gen9_set_dc_state(i915, DC_STATE_DISABLE);
> >   	/* 1. Disable all display engine functions -> aready done */
> >   	/* 2. Disable DBUF */
> > -	icl_dbuf_disable(dev_priv);
> > +	icl_dbuf_disable(i915);
> >   	/* 3. Disable CD clock */
> > -	intel_cdclk_uninit(dev_priv);
> > +	intel_cdclk_uninit(i915);
> >   	/*
> >   	 * 4. Disable Power Well 1 (PG1).
> > @@ -4157,20 +4157,20 @@ void icl_display_core_uninit(struct drm_i915_private *dev_priv)
> >   	 *    disabled at this point.
> >   	 */
> >   	mutex_lock(&power_domains->lock);
> > -	well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
> > -	intel_power_well_disable(dev_priv, well);
> > +	well = lookup_power_well(i915, SKL_DISP_PW_1);
> > +	intel_power_well_disable(i915, well);
> >   	mutex_unlock(&power_domains->lock);
> >   	/* 5. */
> > -	intel_combo_phy_uninit(dev_priv);
> > +	intel_combo_phy_uninit(i915);
> >   }
> > -static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> > +static void chv_phy_control_init(struct drm_i915_private *i915)
> >   {
> >   	struct i915_power_well *cmn_bc =
> > -		lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> > +		lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> >   	struct i915_power_well *cmn_d =
> > -		lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
> > +		lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D);
> >   	/*
> >   	 * DISPLAY_PHY_CONTROL can get corrupted if read. As a
> > @@ -4179,7 +4179,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> >   	 * power well state and lane status to reconstruct the
> >   	 * expected initial value.
> >   	 */
> > -	dev_priv->chv_phy_control =
> > +	i915->chv_phy_control =
> >   		PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY0) |
> >   		PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY1) |
> >   		PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY0, DPIO_CH0) |
> > @@ -4193,7 +4193,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> >   	 * override and set the lane powerdown bits accding to the
> >   	 * current lane status.
> >   	 */
> > -	if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
> > +	if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) {
> >   		u32 status = I915_READ(DPLL(PIPE_A));
> >   		unsigned int mask;
> > @@ -4201,30 +4201,30 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> >   		if (mask == 0xf)
> >   			mask = 0x0;
> >   		else
> > -			dev_priv->chv_phy_control |=
> > +			i915->chv_phy_control |=
> >   				PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH0);
> > -		dev_priv->chv_phy_control |=
> > +		i915->chv_phy_control |=
> >   			PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH0);
> >   		mask = (status & DPLL_PORTC_READY_MASK) >> 4;
> >   		if (mask == 0xf)
> >   			mask = 0x0;
> >   		else
> > -			dev_priv->chv_phy_control |=
> > +			i915->chv_phy_control |=
> >   				PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH1);
> > -		dev_priv->chv_phy_control |=
> > +		i915->chv_phy_control |=
> >   			PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH1);
> > -		dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0);
> > +		i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0);
> > -		dev_priv->chv_phy_assert[DPIO_PHY0] = false;
> > +		i915->chv_phy_assert[DPIO_PHY0] = false;
> >   	} else {
> > -		dev_priv->chv_phy_assert[DPIO_PHY0] = true;
> > +		i915->chv_phy_assert[DPIO_PHY0] = true;
> >   	}
> > -	if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
> > +	if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) {
> >   		u32 status = I915_READ(DPIO_PHY_STATUS);
> >   		unsigned int mask;
> > @@ -4233,42 +4233,42 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv)
> >   		if (mask == 0xf)
> >   			mask = 0x0;
> >   		else
> > -			dev_priv->chv_phy_control |=
> > +			i915->chv_phy_control |=
> >   				PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY1, DPIO_CH0);
> > -		dev_priv->chv_phy_control |=
> > +		i915->chv_phy_control |=
> >   			PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY1, DPIO_CH0);
> > -		dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1);
> > +		i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1);
> > -		dev_priv->chv_phy_assert[DPIO_PHY1] = false;
> > +		i915->chv_phy_assert[DPIO_PHY1] = false;
> >   	} else {
> > -		dev_priv->chv_phy_assert[DPIO_PHY1] = true;
> > +		i915->chv_phy_assert[DPIO_PHY1] = true;
> >   	}
> > -	I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
> > +	I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control);
> >   	DRM_DEBUG_KMS("Initial PHY_CONTROL=0x%08x\n",
> > -		      dev_priv->chv_phy_control);
> > +		      i915->chv_phy_control);
> >   }
> > -static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv)
> > +static void vlv_cmnlane_wa(struct drm_i915_private *i915)
> >   {
> >   	struct i915_power_well *cmn =
> > -		lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
> > +		lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC);
> >   	struct i915_power_well *disp2d =
> > -		lookup_power_well(dev_priv, VLV_DISP_PW_DISP2D);
> > +		lookup_power_well(i915, VLV_DISP_PW_DISP2D);
> >   	/* If the display might be already active skip this */
> > -	if (cmn->desc->ops->is_enabled(dev_priv, cmn) &&
> > -	    disp2d->desc->ops->is_enabled(dev_priv, disp2d) &&
> > +	if (cmn->desc->ops->is_enabled(i915, cmn) &&
> > +	    disp2d->desc->ops->is_enabled(i915, disp2d) &&
> >   	    I915_READ(DPIO_CTL) & DPIO_CMNRST)
> >   		return;
> >   	DRM_DEBUG_KMS("toggling display PHY side reset\n");
> >   	/* cmnlane needs DPLL registers */
> > -	disp2d->desc->ops->enable(dev_priv, disp2d);
> > +	disp2d->desc->ops->enable(i915, disp2d);
> >   	/*
> >   	 * From VLV2A0_DP_eDP_HDMI_DPIO_driver_vbios_notes_11.docx:
> > @@ -4277,27 +4277,27 @@ static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv)
> >   	 * Simply ungating isn't enough to reset the PHY enough to get
> >   	 * ports and lanes running.
> >   	 */
> > -	cmn->desc->ops->disable(dev_priv, cmn);
> > +	cmn->desc->ops->disable(i915, cmn);
> >   }
> > -static bool vlv_punit_is_power_gated(struct drm_i915_private *dev_priv, u32 reg0)
> > +static bool vlv_punit_is_power_gated(struct drm_i915_private *i915, u32 reg0)
> >   {
> >   	bool ret;
> > -	vlv_punit_get(dev_priv);
> > -	ret = (vlv_punit_read(dev_priv, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE;
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_get(i915);
> > +	ret = (vlv_punit_read(i915, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE;
> > +	vlv_punit_put(i915);
> >   	return ret;
> >   }
> > -static void assert_ved_power_gated(struct drm_i915_private *dev_priv)
> > +static void assert_ved_power_gated(struct drm_i915_private *i915)
> >   {
> > -	WARN(!vlv_punit_is_power_gated(dev_priv, PUNIT_REG_VEDSSPM0),
> > +	WARN(!vlv_punit_is_power_gated(i915, PUNIT_REG_VEDSSPM0),
> >   	     "VED not power gated\n");
> >   }
> > -static void assert_isp_power_gated(struct drm_i915_private *dev_priv)
> > +static void assert_isp_power_gated(struct drm_i915_private *i915)
> >   {
> >   	static const struct pci_device_id isp_ids[] = {
> >   		{PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f38)},
> > @@ -4306,11 +4306,11 @@ static void assert_isp_power_gated(struct drm_i915_private *dev_priv)
> >   	};
> >   	WARN(!pci_dev_present(isp_ids) &&
> > -	     !vlv_punit_is_power_gated(dev_priv, PUNIT_REG_ISPSSPM0),
> > +	     !vlv_punit_is_power_gated(i915, PUNIT_REG_ISPSSPM0),
> >   	     "ISP not power gated\n");
> >   }
> > -static void intel_power_domains_verify_state(struct drm_i915_private *dev_priv);
> > +static void intel_power_domains_verify_state(struct drm_i915_private *i915);
> >   /**
> >    * intel_power_domains_init_hw - initialize hardware power domain state
> > diff --git a/drivers/gpu/drm/i915/intel_display_power.h b/drivers/gpu/drm/i915/intel_display_power.h
> > index ff57b0a7fe59..0310fe114ecc 100644
> > --- a/drivers/gpu/drm/i915/intel_display_power.h
> > +++ b/drivers/gpu/drm/i915/intel_display_power.h
> > @@ -80,23 +80,23 @@ struct i915_power_well_ops {
> >   	 * during driver init and resume time, possibly after first calling
> >   	 * the enable/disable handlers.
> >   	 */
> > -	void (*sync_hw)(struct drm_i915_private *dev_priv,
> > +	void (*sync_hw)(struct drm_i915_private *i915,
> >   			struct i915_power_well *power_well);
> >   	/*
> >   	 * Enable the well and resources that depend on it (for example
> >   	 * interrupts located on the well). Called after the 0->1 refcount
> >   	 * transition.
> >   	 */
> > -	void (*enable)(struct drm_i915_private *dev_priv,
> > +	void (*enable)(struct drm_i915_private *i915,
> >   		       struct i915_power_well *power_well);
> >   	/*
> >   	 * Disable the well and resources that depend on it. Called after
> >   	 * the 1->0 refcount transition.
> >   	 */
> > -	void (*disable)(struct drm_i915_private *dev_priv,
> > +	void (*disable)(struct drm_i915_private *i915,
> >   			struct i915_power_well *power_well);
> >   	/* Returns the hw enabled state. */
> > -	bool (*is_enabled)(struct drm_i915_private *dev_priv,
> > +	bool (*is_enabled)(struct drm_i915_private *i915,
> >   			   struct i915_power_well *power_well);
> >   };
> > @@ -204,48 +204,48 @@ struct i915_power_domains {
> >   	for_each_power_well_reverse(__dev_priv, __power_well)		        \
> >   		for_each_if((__power_well)->desc->domains & (__domain_mask))
> > -void skl_enable_dc6(struct drm_i915_private *dev_priv);
> > -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv);
> > -void bxt_enable_dc9(struct drm_i915_private *dev_priv);
> > -void bxt_disable_dc9(struct drm_i915_private *dev_priv);
> > -void gen9_enable_dc5(struct drm_i915_private *dev_priv);
> > +void skl_enable_dc6(struct drm_i915_private *i915);
> > +void gen9_sanitize_dc_state(struct drm_i915_private *i915);
> > +void bxt_enable_dc9(struct drm_i915_private *i915);
> > +void bxt_disable_dc9(struct drm_i915_private *i915);
> > +void gen9_enable_dc5(struct drm_i915_private *i915);
> > -int intel_power_domains_init(struct drm_i915_private *dev_priv);
> > -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv);
> > -void intel_power_domains_init_hw(struct drm_i915_private *dev_priv, bool resume);
> > -void intel_power_domains_fini_hw(struct drm_i915_private *dev_priv);
> > -void icl_display_core_init(struct drm_i915_private *dev_priv, bool resume);
> > -void icl_display_core_uninit(struct drm_i915_private *dev_priv);
> > -void intel_power_domains_enable(struct drm_i915_private *dev_priv);
> > -void intel_power_domains_disable(struct drm_i915_private *dev_priv);
> > -void intel_power_domains_suspend(struct drm_i915_private *dev_priv,
> > +int intel_power_domains_init(struct drm_i915_private *i915);
> > +void intel_power_domains_cleanup(struct drm_i915_private *i915);
> > +void intel_power_domains_init_hw(struct drm_i915_private *i915, bool resume);
> > +void intel_power_domains_fini_hw(struct drm_i915_private *i915);
> > +void icl_display_core_init(struct drm_i915_private *i915, bool resume);
> > +void icl_display_core_uninit(struct drm_i915_private *i915);
> > +void intel_power_domains_enable(struct drm_i915_private *i915);
> > +void intel_power_domains_disable(struct drm_i915_private *i915);
> > +void intel_power_domains_suspend(struct drm_i915_private *i915,
> >   				 enum i915_drm_suspend_mode);
> > -void intel_power_domains_resume(struct drm_i915_private *dev_priv);
> > -void hsw_enable_pc8(struct drm_i915_private *dev_priv);
> > -void hsw_disable_pc8(struct drm_i915_private *dev_priv);
> > -void bxt_display_core_init(struct drm_i915_private *dev_priv, bool resume);
> > -void bxt_display_core_uninit(struct drm_i915_private *dev_priv);
> > +void intel_power_domains_resume(struct drm_i915_private *i915);
> > +void hsw_enable_pc8(struct drm_i915_private *i915);
> > +void hsw_disable_pc8(struct drm_i915_private *i915);
> > +void bxt_display_core_init(struct drm_i915_private *i915, bool resume);
> > +void bxt_display_core_uninit(struct drm_i915_private *i915);
> >   const char *
> >   intel_display_power_domain_str(enum intel_display_power_domain domain);
> > -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_display_power_is_enabled(struct drm_i915_private *i915,
> >   				    enum intel_display_power_domain domain);
> > -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
> > +bool __intel_display_power_is_enabled(struct drm_i915_private *i915,
> >   				      enum intel_display_power_domain domain);
> > -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
> > +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915,
> >   					enum intel_display_power_domain domain);
> >   intel_wakeref_t
> > -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
> > +intel_display_power_get_if_enabled(struct drm_i915_private *i915,
> >   				   enum intel_display_power_domain domain);
> > -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv,
> > +void intel_display_power_put_unchecked(struct drm_i915_private *i915,
> >   				       enum intel_display_power_domain domain);
> >   void __intel_display_power_put_async(struct drm_i915_private *i915,
> >   				     enum intel_display_power_domain domain,
> >   				     intel_wakeref_t wakeref);
> >   void intel_display_power_flush_work(struct drm_i915_private *i915);
> >   #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
> > -void intel_display_power_put(struct drm_i915_private *dev_priv,
> > +void intel_display_power_put(struct drm_i915_private *i915,
> >   			     enum intel_display_power_domain domain,
> >   			     intel_wakeref_t wakeref);
> >   static inline void
> > @@ -277,12 +277,12 @@ intel_display_power_put_async(struct drm_i915_private *i915,
> >   	for ((wf) = intel_display_power_get((i915), (domain)); (wf); \
> >   	     intel_display_power_put_async((i915), (domain), (wf)), (wf) = 0)
> > -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
> > +void icl_dbuf_slices_update(struct drm_i915_private *i915,
> >   			    u8 req_slices);
> >   void chv_phy_powergate_lanes(struct intel_encoder *encoder,
> >   			     bool override, unsigned int mask);
> > -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> > +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy,
> >   			  enum dpio_channel ch, bool override);
> >   #endif /* __INTEL_DISPLAY_POWER_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c
> > index 4336df46fe78..d4dc2b9506e5 100644
> > --- a/drivers/gpu/drm/i915/intel_dp.c
> > +++ b/drivers/gpu/drm/i915/intel_dp.c
> > @@ -160,7 +160,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp);
> >   static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync);
> >   static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> >   					   const struct intel_crtc_state *crtc_state);
> > -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> > +static void vlv_steal_power_sequencer(struct drm_i915_private *i915,
> >   				      enum pipe pipe);
> >   static void intel_dp_unset_edid(struct intel_dp *intel_dp);
> > @@ -214,8 +214,8 @@ static int intel_dp_max_common_rate(struct intel_dp *intel_dp)
> >   static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
> >   	intel_wakeref_t wakeref;
> >   	u32 lane_info;
> > @@ -223,7 +223,7 @@ static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp)
> >   		return 4;
> >   	lane_info = 0;
> > -	with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> > +	with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> >   		lane_info = (I915_READ(PORT_TX_DFLEXDPSP) &
> >   			     DP_LANE_ASSIGNMENT_MASK(tc_port)) >>
> >   				DP_LANE_ASSIGNMENT_SHIFT(tc_port);
> > @@ -284,8 +284,8 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	struct intel_encoder *encoder = &intel_dig_port->base;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > -	int max_dotclk = dev_priv->max_dotclk_freq;
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > +	int max_dotclk = i915->max_dotclk_freq;
> >   	int ds_max_dotclk;
> >   	int type = intel_dp->downstream_ports[0] & DP_DS_PORT_TYPE_MASK;
> > @@ -305,7 +305,7 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp)
> >   static int cnl_max_source_rate(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	enum port port = dig_port->base.port;
> >   	u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK;
> > @@ -315,7 +315,7 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp)
> >   		return 540000;
> >   	/* For this SKU 8.1G is supported in all ports */
> > -	if (IS_CNL_WITH_PORT_F(dev_priv))
> > +	if (IS_CNL_WITH_PORT_F(i915))
> >   		return 810000;
> >   	/* For other SKUs, max rate on ports A and D is 5.4G */
> > @@ -328,11 +328,11 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp)
> >   static int icl_max_source_rate(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	enum port port = dig_port->base.port;
> > -	if (intel_port_is_combophy(dev_priv, port) &&
> > -	    !IS_ELKHARTLAKE(dev_priv) &&
> > +	if (intel_port_is_combophy(i915, port) &&
> > +	    !IS_ELKHARTLAKE(i915) &&
> >   	    !intel_dp_is_edp(intel_dp))
> >   		return 540000;
> > @@ -359,30 +359,30 @@ intel_dp_set_source_rates(struct intel_dp *intel_dp)
> >   		162000, 270000
> >   	};
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	const struct ddi_vbt_port_info *info =
> > -		&dev_priv->vbt.ddi_port_info[dig_port->base.port];
> > +		&i915->vbt.ddi_port_info[dig_port->base.port];
> >   	const int *source_rates;
> >   	int size, max_rate = 0, vbt_max_rate = info->dp_max_link_rate;
> >   	/* This should only be done once */
> >   	WARN_ON(intel_dp->source_rates || intel_dp->num_source_rates);
> > -	if (INTEL_GEN(dev_priv) >= 10) {
> > +	if (INTEL_GEN(i915) >= 10) {
> >   		source_rates = cnl_rates;
> >   		size = ARRAY_SIZE(cnl_rates);
> > -		if (IS_GEN(dev_priv, 10))
> > +		if (IS_GEN(i915, 10))
> >   			max_rate = cnl_max_source_rate(intel_dp);
> >   		else
> >   			max_rate = icl_max_source_rate(intel_dp);
> > -	} else if (IS_GEN9_LP(dev_priv)) {
> > +	} else if (IS_GEN9_LP(i915)) {
> >   		source_rates = bxt_rates;
> >   		size = ARRAY_SIZE(bxt_rates);
> > -	} else if (IS_GEN9_BC(dev_priv)) {
> > +	} else if (IS_GEN9_BC(i915)) {
> >   		source_rates = skl_rates;
> >   		size = ARRAY_SIZE(skl_rates);
> > -	} else if ((IS_HASWELL(dev_priv) && !IS_HSW_ULX(dev_priv)) ||
> > -		   IS_BROADWELL(dev_priv)) {
> > +	} else if ((IS_HASWELL(i915) && !IS_HSW_ULX(i915)) ||
> > +		   IS_BROADWELL(i915)) {
> >   		source_rates = hsw_rates;
> >   		size = ARRAY_SIZE(hsw_rates);
> >   	} else {
> > @@ -532,7 +532,7 @@ intel_dp_mode_valid(struct drm_connector *connector,
> >   	struct intel_dp *intel_dp = intel_attached_dp(connector);
> >   	struct intel_connector *intel_connector = to_intel_connector(connector);
> >   	struct drm_display_mode *fixed_mode = intel_connector->panel.fixed_mode;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	int target_clock = mode->clock;
> >   	int max_rate, mode_rate, max_lanes, max_link_clock;
> >   	int max_dotclk;
> > @@ -564,7 +564,7 @@ intel_dp_mode_valid(struct drm_connector *connector,
> >   	 * Output bpp is stored in 6.4 format so right shift by 4 to get the
> >   	 * integer value since we support only integer values of bpp.
> >   	 */
> > -	if ((INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) &&
> > +	if ((INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) &&
> >   	    drm_dp_sink_supports_dsc(intel_dp->dsc_dpcd)) {
> >   		if (intel_dp_is_edp(intel_dp)) {
> >   			dsc_max_output_bpp =
> > @@ -630,17 +630,17 @@ intel_dp_pps_init(struct intel_dp *intel_dp);
> >   static intel_wakeref_t
> >   pps_lock(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	intel_wakeref_t wakeref;
> >   	/*
> >   	 * See intel_power_sequencer_reset() why we need
> >   	 * a power domain reference here.
> >   	 */
> > -	wakeref = intel_display_power_get(dev_priv,
> > +	wakeref = intel_display_power_get(i915,
> >   					  intel_aux_power_domain(dp_to_dig_port(intel_dp)));
> > -	mutex_lock(&dev_priv->pps_mutex);
> > +	mutex_lock(&i915->pps_mutex);
> >   	return wakeref;
> >   }
> > @@ -648,10 +648,10 @@ pps_lock(struct intel_dp *intel_dp)
> >   static intel_wakeref_t
> >   pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	mutex_unlock(&dev_priv->pps_mutex);
> > -	intel_display_power_put(dev_priv,
> > +	mutex_unlock(&i915->pps_mutex);
> > +	intel_display_power_put(i915,
> >   				intel_aux_power_domain(dp_to_dig_port(intel_dp)),
> >   				wakeref);
> >   	return 0;
> > @@ -663,7 +663,7 @@ pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref)
> >   static void
> >   vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum pipe pipe = intel_dp->pps_pipe;
> >   	bool pll_enabled, release_cl_override = false;
> > @@ -687,7 +687,7 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> >   	DP |= DP_PORT_WIDTH(1);
> >   	DP |= DP_LINK_TRAIN_PAT_1;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > +	if (IS_CHERRYVIEW(i915))
> >   		DP |= DP_PIPE_SEL_CHV(pipe);
> >   	else
> >   		DP |= DP_PIPE_SEL(pipe);
> > @@ -699,10 +699,10 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> >   	 * So enable temporarily it if it's not already enabled.
> >   	 */
> >   	if (!pll_enabled) {
> > -		release_cl_override = IS_CHERRYVIEW(dev_priv) &&
> > -			!chv_phy_powergate_ch(dev_priv, phy, ch, true);
> > +		release_cl_override = IS_CHERRYVIEW(i915) &&
> > +			!chv_phy_powergate_ch(i915, phy, ch, true);
> > -		if (vlv_force_pll_on(dev_priv, pipe, IS_CHERRYVIEW(dev_priv) ?
> > +		if (vlv_force_pll_on(i915, pipe, IS_CHERRYVIEW(i915) ?
> >   				     &chv_dpll[0].dpll : &vlv_dpll[0].dpll)) {
> >   			DRM_ERROR("Failed to force on pll for pipe %c!\n",
> >   				  pipe_name(pipe));
> > @@ -726,14 +726,14 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp)
> >   	POSTING_READ(intel_dp->output_reg);
> >   	if (!pll_enabled) {
> > -		vlv_force_pll_off(dev_priv, pipe);
> > +		vlv_force_pll_off(i915, pipe);
> >   		if (release_cl_override)
> > -			chv_phy_powergate_ch(dev_priv, phy, ch, false);
> > +			chv_phy_powergate_ch(i915, phy, ch, false);
> >   	}
> >   }
> > -static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> > +static enum pipe vlv_find_free_pps(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> >   	unsigned int pipes = (1 << PIPE_A) | (1 << PIPE_B);
> > @@ -742,7 +742,7 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> >   	 * We don't have power sequencer currently.
> >   	 * Pick one that's not used by other ports.
> >   	 */
> > -	for_each_intel_dp(&dev_priv->drm, encoder) {
> > +	for_each_intel_dp(&i915->drm, encoder) {
> >   		struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   		if (encoder->type == INTEL_OUTPUT_EDP) {
> > @@ -768,11 +768,11 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv)
> >   static enum pipe
> >   vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum pipe pipe;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	/* We should never land here with regular DP ports */
> >   	WARN_ON(!intel_dp_is_edp(intel_dp));
> > @@ -783,7 +783,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> >   	if (intel_dp->pps_pipe != INVALID_PIPE)
> >   		return intel_dp->pps_pipe;
> > -	pipe = vlv_find_free_pps(dev_priv);
> > +	pipe = vlv_find_free_pps(i915);
> >   	/*
> >   	 * Didn't find one. This should not happen since there
> > @@ -792,7 +792,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> >   	if (WARN_ON(pipe == INVALID_PIPE))
> >   		pipe = PIPE_A;
> > -	vlv_steal_power_sequencer(dev_priv, pipe);
> > +	vlv_steal_power_sequencer(i915, pipe);
> >   	intel_dp->pps_pipe = pipe;
> >   	DRM_DEBUG_KMS("picked pipe %c power sequencer for port %c\n",
> > @@ -815,10 +815,10 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp)
> >   static int
> >   bxt_power_sequencer_idx(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > -	int backlight_controller = dev_priv->vbt.backlight.controller;
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > +	int backlight_controller = i915->vbt.backlight.controller;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	/* We should never land here with regular DP ports */
> >   	WARN_ON(!intel_dp_is_edp(intel_dp));
> > @@ -837,29 +837,29 @@ bxt_power_sequencer_idx(struct intel_dp *intel_dp)
> >   	return backlight_controller;
> >   }
> > -typedef bool (*vlv_pipe_check)(struct drm_i915_private *dev_priv,
> > +typedef bool (*vlv_pipe_check)(struct drm_i915_private *i915,
> >   			       enum pipe pipe);
> > -static bool vlv_pipe_has_pp_on(struct drm_i915_private *dev_priv,
> > +static bool vlv_pipe_has_pp_on(struct drm_i915_private *i915,
> >   			       enum pipe pipe)
> >   {
> >   	return I915_READ(PP_STATUS(pipe)) & PP_ON;
> >   }
> > -static bool vlv_pipe_has_vdd_on(struct drm_i915_private *dev_priv,
> > +static bool vlv_pipe_has_vdd_on(struct drm_i915_private *i915,
> >   				enum pipe pipe)
> >   {
> >   	return I915_READ(PP_CONTROL(pipe)) & EDP_FORCE_VDD;
> >   }
> > -static bool vlv_pipe_any(struct drm_i915_private *dev_priv,
> > +static bool vlv_pipe_any(struct drm_i915_private *i915,
> >   			 enum pipe pipe)
> >   {
> >   	return true;
> >   }
> >   static enum pipe
> > -vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> > +vlv_initial_pps_pipe(struct drm_i915_private *i915,
> >   		     enum port port,
> >   		     vlv_pipe_check pipe_check)
> >   {
> > @@ -872,7 +872,7 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> >   		if (port_sel != PANEL_PORT_SELECT_VLV(port))
> >   			continue;
> > -		if (!pipe_check(dev_priv, pipe))
> > +		if (!pipe_check(i915, pipe))
> >   			continue;
> >   		return pipe;
> > @@ -884,23 +884,23 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv,
> >   static void
> >   vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum port port = intel_dig_port->base.port;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	/* try to find a pipe with this port selected */
> >   	/* first pick one where the panel is on */
> > -	intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> > +	intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> >   						  vlv_pipe_has_pp_on);
> >   	/* didn't find one? pick one where vdd is on */
> >   	if (intel_dp->pps_pipe == INVALID_PIPE)
> > -		intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> > +		intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> >   							  vlv_pipe_has_vdd_on);
> >   	/* didn't find one? pick one with just the correct port */
> >   	if (intel_dp->pps_pipe == INVALID_PIPE)
> > -		intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port,
> > +		intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port,
> >   							  vlv_pipe_any);
> >   	/* didn't find one? just let vlv_power_sequencer_pipe() pick one when needed */
> > @@ -917,12 +917,12 @@ vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp)
> >   	intel_dp_init_panel_power_sequencer_registers(intel_dp, false);
> >   }
> > -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> > +void intel_power_sequencer_reset(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> > -	if (WARN_ON(!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> > -		    !IS_GEN9_LP(dev_priv)))
> > +	if (WARN_ON(!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> > +		    !IS_GEN9_LP(i915)))
> >   		return;
> >   	/*
> > @@ -935,7 +935,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> >   	 * should use them always.
> >   	 */
> > -	for_each_intel_dp(&dev_priv->drm, encoder) {
> > +	for_each_intel_dp(&i915->drm, encoder) {
> >   		struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   		WARN_ON(intel_dp->active_pipe != INVALID_PIPE);
> > @@ -943,7 +943,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv)
> >   		if (encoder->type != INTEL_OUTPUT_EDP)
> >   			continue;
> > -		if (IS_GEN9_LP(dev_priv))
> > +		if (IS_GEN9_LP(i915))
> >   			intel_dp->pps_reset = true;
> >   		else
> >   			intel_dp->pps_pipe = INVALID_PIPE;
> > @@ -961,14 +961,14 @@ struct pps_registers {
> >   static void intel_pps_get_registers(struct intel_dp *intel_dp,
> >   				    struct pps_registers *regs)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	int pps_idx = 0;
> >   	memset(regs, 0, sizeof(*regs));
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		pps_idx = bxt_power_sequencer_idx(intel_dp);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		pps_idx = vlv_power_sequencer_pipe(intel_dp);
> >   	regs->pp_ctrl = PP_CONTROL(pps_idx);
> > @@ -977,7 +977,7 @@ static void intel_pps_get_registers(struct intel_dp *intel_dp,
> >   	regs->pp_off = PP_OFF_DELAYS(pps_idx);
> >   	/* Cycle delay moved from PP_DIVISOR to PP_CONTROL */
> > -	if (IS_GEN9_LP(dev_priv) || INTEL_PCH_TYPE(dev_priv) >= PCH_CNP)
> > +	if (IS_GEN9_LP(i915) || INTEL_PCH_TYPE(i915) >= PCH_CNP)
> >   		regs->pp_div = INVALID_MMIO_REG;
> >   	else
> >   		regs->pp_div = PP_DIVISOR(pps_idx);
> > @@ -1010,14 +1010,14 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code,
> >   {
> >   	struct intel_dp *intel_dp = container_of(this, typeof(* intel_dp),
> >   						 edp_notifier);
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	intel_wakeref_t wakeref;
> >   	if (!intel_dp_is_edp(intel_dp) || code != SYS_RESTART)
> >   		return 0;
> >   	with_pps_lock(intel_dp, wakeref) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   			enum pipe pipe = vlv_power_sequencer_pipe(intel_dp);
> >   			i915_reg_t pp_ctrl_reg, pp_div_reg;
> >   			u32 pp_div;
> > @@ -1039,11 +1039,11 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code,
> >   static bool edp_have_panel_power(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   	    intel_dp->pps_pipe == INVALID_PIPE)
> >   		return false;
> > @@ -1052,11 +1052,11 @@ static bool edp_have_panel_power(struct intel_dp *intel_dp)
> >   static bool edp_have_panel_vdd(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   	    intel_dp->pps_pipe == INVALID_PIPE)
> >   		return false;
> > @@ -1066,7 +1066,7 @@ static bool edp_have_panel_vdd(struct intel_dp *intel_dp)
> >   static void
> >   intel_dp_check_edp(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (!intel_dp_is_edp(intel_dp))
> >   		return;
> > @@ -1103,7 +1103,7 @@ intel_dp_aux_wait_done(struct intel_dp *intel_dp)
> >   static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (index)
> >   		return 0;
> > @@ -1112,12 +1112,12 @@ static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> >   	 * The clock divider is based off the hrawclk, and would like to run at
> >   	 * 2MHz.  So, take the hrawclk value and divide by 2000 and use that
> >   	 */
> > -	return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000);
> > +	return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000);
> >   }
> >   static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	if (index)
> > @@ -1129,17 +1129,17 @@ static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> >   	 * divide by 2000 and use that
> >   	 */
> >   	if (dig_port->aux_ch == AUX_CH_A)
> > -		return DIV_ROUND_CLOSEST(dev_priv->cdclk.hw.cdclk, 2000);
> > +		return DIV_ROUND_CLOSEST(i915->cdclk.hw.cdclk, 2000);
> >   	else
> > -		return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000);
> > +		return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000);
> >   }
> >   static u32 hsw_get_aux_clock_divider(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(dev_priv)) {
> > +	if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(i915)) {
> >   		/* Workaround for non-ULT HSW */
> >   		switch (index) {
> >   		case 0: return 63;
> > @@ -1166,16 +1166,16 @@ static u32 g4x_get_aux_send_ctl(struct intel_dp *intel_dp,
> >   				u32 aux_clock_divider)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   			to_i915(intel_dig_port->base.base.dev);
> >   	u32 precharge, timeout;
> > -	if (IS_GEN(dev_priv, 6))
> > +	if (IS_GEN(i915, 6))
> >   		precharge = 3;
> >   	else
> >   		precharge = 5;
> > -	if (IS_BROADWELL(dev_priv))
> > +	if (IS_BROADWELL(i915))
> >   		timeout = DP_AUX_CH_CTL_TIME_OUT_600us;
> >   	else
> >   		timeout = DP_AUX_CH_CTL_TIME_OUT_400us;
> > @@ -1481,7 +1481,7 @@ intel_dp_aux_transfer(struct drm_dp_aux *aux, struct drm_dp_aux_msg *msg)
> >   static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1498,7 +1498,7 @@ static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp)
> >   static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1515,7 +1515,7 @@ static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index)
> >   static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1534,7 +1534,7 @@ static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp)
> >   static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1553,7 +1553,7 @@ static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index)
> >   static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1573,7 +1573,7 @@ static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp)
> >   static i915_reg_t skl_aux_data_reg(struct intel_dp *intel_dp, int index)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	enum aux_ch aux_ch = dig_port->aux_ch;
> > @@ -1600,14 +1600,14 @@ intel_dp_aux_fini(struct intel_dp *intel_dp)
> >   static void
> >   intel_dp_aux_init(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	struct intel_encoder *encoder = &dig_port->base;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		intel_dp->aux_ch_ctl_reg = skl_aux_ctl_reg;
> >   		intel_dp->aux_ch_data_reg = skl_aux_data_reg;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		intel_dp->aux_ch_ctl_reg = ilk_aux_ctl_reg;
> >   		intel_dp->aux_ch_data_reg = ilk_aux_data_reg;
> >   	} else {
> > @@ -1615,16 +1615,16 @@ intel_dp_aux_init(struct intel_dp *intel_dp)
> >   		intel_dp->aux_ch_data_reg = g4x_aux_data_reg;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		intel_dp->get_aux_clock_divider = skl_get_aux_clock_divider;
> > -	else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv))
> > +	else if (IS_BROADWELL(i915) || IS_HASWELL(i915))
> >   		intel_dp->get_aux_clock_divider = hsw_get_aux_clock_divider;
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > +	else if (HAS_PCH_SPLIT(i915))
> >   		intel_dp->get_aux_clock_divider = ilk_get_aux_clock_divider;
> >   	else
> >   		intel_dp->get_aux_clock_divider = g4x_get_aux_clock_divider;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		intel_dp->get_aux_send_ctl = skl_get_aux_send_ctl;
> >   	else
> >   		intel_dp->get_aux_send_ctl = g4x_get_aux_send_ctl;
> > @@ -1655,20 +1655,20 @@ static void
> >   intel_dp_set_clock(struct intel_encoder *encoder,
> >   		   struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	const struct dp_link_dpll *divisor = NULL;
> >   	int i, count = 0;
> > -	if (IS_G4X(dev_priv)) {
> > +	if (IS_G4X(i915)) {
> >   		divisor = g4x_dpll;
> >   		count = ARRAY_SIZE(g4x_dpll);
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		divisor = pch_dpll;
> >   		count = ARRAY_SIZE(pch_dpll);
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_CHERRYVIEW(i915)) {
> >   		divisor = chv_dpll;
> >   		count = ARRAY_SIZE(chv_dpll);
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		divisor = vlv_dpll;
> >   		count = ARRAY_SIZE(vlv_dpll);
> >   	}
> > @@ -1760,9 +1760,9 @@ void intel_dp_compute_rate(struct intel_dp *intel_dp, int port_clock,
> >   static bool intel_dp_source_supports_fec(struct intel_dp *intel_dp,
> >   					 const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	return INTEL_GEN(dev_priv) >= 11 &&
> > +	return INTEL_GEN(i915) >= 11 &&
> >   		pipe_config->cpu_transcoder != TRANSCODER_A;
> >   }
> > @@ -1776,9 +1776,9 @@ static bool intel_dp_supports_fec(struct intel_dp *intel_dp,
> >   static bool intel_dp_source_supports_dsc(struct intel_dp *intel_dp,
> >   					 const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	return INTEL_GEN(dev_priv) >= 10 &&
> > +	return INTEL_GEN(i915) >= 10 &&
> >   		pipe_config->cpu_transcoder != TRANSCODER_A;
> >   }
> > @@ -1795,7 +1795,7 @@ static bool intel_dp_supports_dsc(struct intel_dp *intel_dp,
> >   static int intel_dp_compute_bpp(struct intel_dp *intel_dp,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_connector *intel_connector = intel_dp->attached_connector;
> >   	int bpp, bpc;
> > @@ -1808,10 +1808,10 @@ static int intel_dp_compute_bpp(struct intel_dp *intel_dp,
> >   	if (intel_dp_is_edp(intel_dp)) {
> >   		/* Get bpp from vbt only for panels that dont have bpp in edid */
> >   		if (intel_connector->base.display_info.bpc == 0 &&
> > -		    dev_priv->vbt.edp.bpp && dev_priv->vbt.edp.bpp < bpp) {
> > +		    i915->vbt.edp.bpp && i915->vbt.edp.bpp < bpp) {
> >   			DRM_DEBUG_KMS("clamping bpp for eDP panel to BIOS-provided %i\n",
> > -				      dev_priv->vbt.edp.bpp);
> > -			bpp = dev_priv->vbt.edp.bpp;
> > +				      i915->vbt.edp.bpp);
> > +			bpp = i915->vbt.edp.bpp;
> >   		}
> >   	}
> > @@ -1924,7 +1924,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp,
> >   				       struct link_config_limits *limits)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> >   	u8 dsc_max_bpc;
> >   	int pipe_bpp;
> > @@ -1988,7 +1988,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp,
> >   	 * is greater than the maximum Cdclock and if slice count is even
> >   	 * then we need to use 2 VDSC instances.
> >   	 */
> > -	if (adjusted_mode->crtc_clock > dev_priv->max_cdclk_freq) {
> > +	if (adjusted_mode->crtc_clock > i915->max_cdclk_freq) {
> >   		if (pipe_config->dsc_params.slice_count > 1) {
> >   			pipe_config->dsc_params.dsc_split = true;
> >   		} else {
> > @@ -2168,7 +2168,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   			struct intel_crtc_state *pipe_config,
> >   			struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_lspcon *lspcon = enc_to_intel_lspcon(&encoder->base);
> > @@ -2181,7 +2181,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   					   DP_DPCD_QUIRK_CONSTANT_N);
> >   	int ret = 0, output_bpp;
> > -	if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv) && port != PORT_A)
> > +	if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915) && port != PORT_A)
> >   		pipe_config->has_pch_encoder = true;
> >   	pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB;
> > @@ -2195,7 +2195,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   		return ret;
> >   	pipe_config->has_drrs = false;
> > -	if (IS_G4X(dev_priv) || port == PORT_A)
> > +	if (IS_G4X(i915) || port == PORT_A)
> >   		pipe_config->has_audio = false;
> >   	else if (intel_conn_state->force_audio == HDMI_AUDIO_AUTO)
> >   		pipe_config->has_audio = intel_dp->has_audio;
> > @@ -2206,13 +2206,13 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   		intel_fixed_panel_mode(intel_connector->panel.fixed_mode,
> >   				       adjusted_mode);
> > -		if (INTEL_GEN(dev_priv) >= 9) {
> > +		if (INTEL_GEN(i915) >= 9) {
> >   			ret = skl_update_scaler_crtc(pipe_config);
> >   			if (ret)
> >   				return ret;
> >   		}
> > -		if (HAS_GMCH(dev_priv))
> > +		if (HAS_GMCH(i915))
> >   			intel_gmch_panel_fitting(intel_crtc, pipe_config,
> >   						 conn_state->scaling_mode);
> >   		else
> > @@ -2223,7 +2223,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   	if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN)
> >   		return -EINVAL;
> > -	if (HAS_GMCH(dev_priv) &&
> > +	if (HAS_GMCH(i915) &&
> >   	    adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE)
> >   		return -EINVAL;
> > @@ -2250,7 +2250,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   			       constant_n);
> >   	if (intel_connector->panel.downclock_mode != NULL &&
> > -		dev_priv->drrs.type == SEAMLESS_DRRS_SUPPORT) {
> > +		i915->drrs.type == SEAMLESS_DRRS_SUPPORT) {
> >   			pipe_config->has_drrs = true;
> >   			intel_link_compute_m_n(output_bpp,
> >   					       pipe_config->lane_count,
> > @@ -2260,7 +2260,7 @@ intel_dp_compute_config(struct intel_encoder *encoder,
> >   					       constant_n);
> >   	}
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		intel_dp_set_clock(encoder, pipe_config);
> >   	intel_psr_compute_config(intel_dp, pipe_config);
> > @@ -2281,7 +2281,7 @@ void intel_dp_set_link_params(struct intel_dp *intel_dp,
> >   static void intel_dp_prepare(struct intel_encoder *encoder,
> >   			     const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	enum port port = encoder->port;
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> > @@ -2320,7 +2320,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> >   	/* Split out the IBX/CPU vs CPT settings */
> > -	if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> > +	if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> >   		if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC)
> >   			intel_dp->DP |= DP_SYNC_HS_HIGH;
> >   		if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC)
> > @@ -2331,7 +2331,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> >   			intel_dp->DP |= DP_ENHANCED_FRAMING;
> >   		intel_dp->DP |= DP_PIPE_SEL_IVB(crtc->pipe);
> > -	} else if (HAS_PCH_CPT(dev_priv) && port != PORT_A) {
> > +	} else if (HAS_PCH_CPT(i915) && port != PORT_A) {
> >   		u32 trans_dp;
> >   		intel_dp->DP |= DP_LINK_TRAIN_OFF_CPT;
> > @@ -2343,7 +2343,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> >   			trans_dp &= ~TRANS_DP_ENH_FRAMING;
> >   		I915_WRITE(TRANS_DP_CTL(crtc->pipe), trans_dp);
> >   	} else {
> > -		if (IS_G4X(dev_priv) && pipe_config->limited_color_range)
> > +		if (IS_G4X(i915) && pipe_config->limited_color_range)
> >   			intel_dp->DP |= DP_COLOR_RANGE_16_235;
> >   		if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC)
> > @@ -2355,7 +2355,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder,
> >   		if (drm_dp_enhanced_frame_cap(intel_dp->dpcd))
> >   			intel_dp->DP |= DP_ENHANCED_FRAMING;
> > -		if (IS_CHERRYVIEW(dev_priv))
> > +		if (IS_CHERRYVIEW(i915))
> >   			intel_dp->DP |= DP_PIPE_SEL_CHV(crtc->pipe);
> >   		else
> >   			intel_dp->DP |= DP_PIPE_SEL(crtc->pipe);
> > @@ -2377,10 +2377,10 @@ static void wait_panel_status(struct intel_dp *intel_dp,
> >   				       u32 mask,
> >   				       u32 value)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	i915_reg_t pp_stat_reg, pp_ctrl_reg;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	intel_pps_verify_state(intel_dp);
> > @@ -2392,7 +2392,7 @@ static void wait_panel_status(struct intel_dp *intel_dp,
> >   			I915_READ(pp_stat_reg),
> >   			I915_READ(pp_ctrl_reg));
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    pp_stat_reg, mask, value,
> >   				    5000))
> >   		DRM_ERROR("Panel status timeout: status %08x control %08x\n",
> > @@ -2453,13 +2453,13 @@ static void edp_wait_backlight_off(struct intel_dp *intel_dp)
> >   static  u32 ironlake_get_pp_control(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 control;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	control = I915_READ(_pp_ctrl_reg(intel_dp));
> > -	if (WARN_ON(!HAS_DDI(dev_priv) &&
> > +	if (WARN_ON(!HAS_DDI(i915) &&
> >   		    (control & PANEL_UNLOCK_MASK) != PANEL_UNLOCK_REGS)) {
> >   		control &= ~PANEL_UNLOCK_MASK;
> >   		control |= PANEL_UNLOCK_REGS;
> > @@ -2474,13 +2474,13 @@ static  u32 ironlake_get_pp_control(struct intel_dp *intel_dp)
> >    */
> >   static bool edp_panel_vdd_on(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	u32 pp;
> >   	i915_reg_t pp_stat_reg, pp_ctrl_reg;
> >   	bool need_to_disable = !intel_dp->want_panel_vdd;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	if (!intel_dp_is_edp(intel_dp))
> >   		return false;
> > @@ -2491,7 +2491,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp)
> >   	if (edp_have_panel_vdd(intel_dp))
> >   		return need_to_disable;
> > -	intel_display_power_get(dev_priv,
> > +	intel_display_power_get(i915,
> >   				intel_aux_power_domain(intel_dig_port));
> >   	DRM_DEBUG_KMS("Turning eDP port %c VDD on\n",
> > @@ -2546,13 +2546,13 @@ void intel_edp_panel_vdd_on(struct intel_dp *intel_dp)
> >   static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port =
> >   		dp_to_dig_port(intel_dp);
> >   	u32 pp;
> >   	i915_reg_t pp_stat_reg, pp_ctrl_reg;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	WARN_ON(intel_dp->want_panel_vdd);
> > @@ -2578,7 +2578,7 @@ static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp)
> >   	if ((pp & PANEL_POWER_ON) == 0)
> >   		intel_dp->panel_power_off_time = ktime_get_boottime();
> > -	intel_display_power_put_unchecked(dev_priv,
> > +	intel_display_power_put_unchecked(i915,
> >   					  intel_aux_power_domain(intel_dig_port));
> >   }
> > @@ -2615,9 +2615,9 @@ static void edp_panel_vdd_schedule_off(struct intel_dp *intel_dp)
> >    */
> >   static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	if (!intel_dp_is_edp(intel_dp))
> >   		return;
> > @@ -2635,11 +2635,11 @@ static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync)
> >   static void edp_panel_on(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 pp;
> >   	i915_reg_t pp_ctrl_reg;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	if (!intel_dp_is_edp(intel_dp))
> >   		return;
> > @@ -2656,7 +2656,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
> >   	pp_ctrl_reg = _pp_ctrl_reg(intel_dp);
> >   	pp = ironlake_get_pp_control(intel_dp);
> > -	if (IS_GEN(dev_priv, 5)) {
> > +	if (IS_GEN(i915, 5)) {
> >   		/* ILK workaround: disable reset around power sequence */
> >   		pp &= ~PANEL_POWER_RESET;
> >   		I915_WRITE(pp_ctrl_reg, pp);
> > @@ -2664,7 +2664,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
> >   	}
> >   	pp |= PANEL_POWER_ON;
> > -	if (!IS_GEN(dev_priv, 5))
> > +	if (!IS_GEN(i915, 5))
> >   		pp |= PANEL_POWER_RESET;
> >   	I915_WRITE(pp_ctrl_reg, pp);
> > @@ -2673,7 +2673,7 @@ static void edp_panel_on(struct intel_dp *intel_dp)
> >   	wait_panel_on(intel_dp);
> >   	intel_dp->last_power_on = jiffies;
> > -	if (IS_GEN(dev_priv, 5)) {
> > +	if (IS_GEN(i915, 5)) {
> >   		pp |= PANEL_POWER_RESET; /* restore panel reset bit */
> >   		I915_WRITE(pp_ctrl_reg, pp);
> >   		POSTING_READ(pp_ctrl_reg);
> > @@ -2694,12 +2694,12 @@ void intel_edp_panel_on(struct intel_dp *intel_dp)
> >   static void edp_panel_off(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	u32 pp;
> >   	i915_reg_t pp_ctrl_reg;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	if (!intel_dp_is_edp(intel_dp))
> >   		return;
> > @@ -2727,7 +2727,7 @@ static void edp_panel_off(struct intel_dp *intel_dp)
> >   	intel_dp->panel_power_off_time = ktime_get_boottime();
> >   	/* We got a reference when we enabled the VDD. */
> > -	intel_display_power_put_unchecked(dev_priv, intel_aux_power_domain(dig_port));
> > +	intel_display_power_put_unchecked(i915, intel_aux_power_domain(dig_port));
> >   }
> >   void intel_edp_panel_off(struct intel_dp *intel_dp)
> > @@ -2744,7 +2744,7 @@ void intel_edp_panel_off(struct intel_dp *intel_dp)
> >   /* Enable backlight in the panel power control. */
> >   static void _intel_edp_backlight_on(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	intel_wakeref_t wakeref;
> >   	/*
> > @@ -2785,7 +2785,7 @@ void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state,
> >   /* Disable backlight in the panel power control. */
> >   static void _intel_edp_backlight_off(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	intel_wakeref_t wakeref;
> >   	if (!intel_dp_is_edp(intel_dp))
> > @@ -2849,7 +2849,7 @@ static void intel_edp_backlight_power(struct intel_connector *connector,
> >   static void assert_dp_port(struct intel_dp *intel_dp, bool state)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev);
> >   	bool cur_state = I915_READ(intel_dp->output_reg) & DP_PORT_EN;
> >   	I915_STATE_WARN(cur_state != state,
> > @@ -2859,7 +2859,7 @@ static void assert_dp_port(struct intel_dp *intel_dp, bool state)
> >   }
> >   #define assert_dp_port_disabled(d) assert_dp_port((d), false)
> > -static void assert_edp_pll(struct drm_i915_private *dev_priv, bool state)
> > +static void assert_edp_pll(struct drm_i915_private *i915, bool state)
> >   {
> >   	bool cur_state = I915_READ(DP_A) & DP_PLL_ENABLE;
> > @@ -2874,11 +2874,11 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp,
> >   				const struct intel_crtc_state *pipe_config)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	assert_pipe_disabled(dev_priv, crtc->pipe);
> > +	assert_pipe_disabled(i915, crtc->pipe);
> >   	assert_dp_port_disabled(intel_dp);
> > -	assert_edp_pll_disabled(dev_priv);
> > +	assert_edp_pll_disabled(i915);
> >   	DRM_DEBUG_KMS("enabling eDP PLL for clock %d\n",
> >   		      pipe_config->port_clock);
> > @@ -2900,8 +2900,8 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp,
> >   	 * 1. Wait for the start of vertical blank on the enabled pipe going to FDI
> >   	 * 2. Program DP PLL enable
> >   	 */
> > -	if (IS_GEN(dev_priv, 5))
> > -		intel_wait_for_vblank_if_active(dev_priv, !crtc->pipe);
> > +	if (IS_GEN(i915, 5))
> > +		intel_wait_for_vblank_if_active(i915, !crtc->pipe);
> >   	intel_dp->DP |= DP_PLL_ENABLE;
> > @@ -2914,11 +2914,11 @@ static void ironlake_edp_pll_off(struct intel_dp *intel_dp,
> >   				 const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	assert_pipe_disabled(dev_priv, crtc->pipe);
> > +	assert_pipe_disabled(i915, crtc->pipe);
> >   	assert_dp_port_disabled(intel_dp);
> > -	assert_edp_pll_enabled(dev_priv);
> > +	assert_edp_pll_enabled(i915);
> >   	DRM_DEBUG_KMS("disabling eDP PLL\n");
> > @@ -2999,12 +2999,12 @@ void intel_dp_sink_dpms(struct intel_dp *intel_dp, int mode)
> >   			      mode == DRM_MODE_DPMS_ON ? "enable" : "disable");
> >   }
> > -static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv,
> > +static bool cpt_dp_port_selected(struct drm_i915_private *i915,
> >   				 enum port port, enum pipe *pipe)
> >   {
> >   	enum pipe p;
> > -	for_each_pipe(dev_priv, p) {
> > +	for_each_pipe(i915, p) {
> >   		u32 val = I915_READ(TRANS_DP_CTL(p));
> >   		if ((val & TRANS_DP_PORT_SEL_MASK) == TRANS_DP_PORT_SEL(port)) {
> > @@ -3021,7 +3021,7 @@ static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv,
> >   	return false;
> >   }
> > -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_dp_port_enabled(struct drm_i915_private *i915,
> >   			   i915_reg_t dp_reg, enum port port,
> >   			   enum pipe *pipe)
> >   {
> > @@ -3033,11 +3033,11 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> >   	ret = val & DP_PORT_EN;
> >   	/* asserts want to know the pipe even if the port is disabled */
> > -	if (IS_IVYBRIDGE(dev_priv) && port == PORT_A)
> > +	if (IS_IVYBRIDGE(i915) && port == PORT_A)
> >   		*pipe = (val & DP_PIPE_SEL_MASK_IVB) >> DP_PIPE_SEL_SHIFT_IVB;
> > -	else if (HAS_PCH_CPT(dev_priv) && port != PORT_A)
> > -		ret &= cpt_dp_port_selected(dev_priv, port, pipe);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (HAS_PCH_CPT(i915) && port != PORT_A)
> > +		ret &= cpt_dp_port_selected(i915, port, pipe);
> > +	else if (IS_CHERRYVIEW(i915))
> >   		*pipe = (val & DP_PIPE_SEL_MASK_CHV) >> DP_PIPE_SEL_SHIFT_CHV;
> >   	else
> >   		*pipe = (val & DP_PIPE_SEL_MASK) >> DP_PIPE_SEL_SHIFT;
> > @@ -3048,20 +3048,20 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> >   static bool intel_dp_get_hw_state(struct intel_encoder *encoder,
> >   				  enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > -	ret = intel_dp_port_enabled(dev_priv, intel_dp->output_reg,
> > +	ret = intel_dp_port_enabled(i915, intel_dp->output_reg,
> >   				    encoder->port, pipe);
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -3069,7 +3069,7 @@ static bool intel_dp_get_hw_state(struct intel_encoder *encoder,
> >   static void intel_dp_get_config(struct intel_encoder *encoder,
> >   				struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	u32 tmp, flags = 0;
> >   	enum port port = encoder->port;
> > @@ -3084,7 +3084,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> >   	pipe_config->has_audio = tmp & DP_AUDIO_OUTPUT_ENABLE && port != PORT_A;
> > -	if (HAS_PCH_CPT(dev_priv) && port != PORT_A) {
> > +	if (HAS_PCH_CPT(i915) && port != PORT_A) {
> >   		u32 trans_dp = I915_READ(TRANS_DP_CTL(crtc->pipe));
> >   		if (trans_dp & TRANS_DP_HSYNC_ACTIVE_HIGH)
> > @@ -3110,7 +3110,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> >   	pipe_config->base.adjusted_mode.flags |= flags;
> > -	if (IS_G4X(dev_priv) && tmp & DP_COLOR_RANGE_16_235)
> > +	if (IS_G4X(i915) && tmp & DP_COLOR_RANGE_16_235)
> >   		pipe_config->limited_color_range = true;
> >   	pipe_config->lane_count =
> > @@ -3129,8 +3129,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> >   		intel_dotclock_calculate(pipe_config->port_clock,
> >   					 &pipe_config->dp_m_n);
> > -	if (intel_dp_is_edp(intel_dp) && dev_priv->vbt.edp.bpp &&
> > -	    pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) {
> > +	if (intel_dp_is_edp(intel_dp) && i915->vbt.edp.bpp &&
> > +	    pipe_config->pipe_bpp > i915->vbt.edp.bpp) {
> >   		/*
> >   		 * This is a big fat ugly hack.
> >   		 *
> > @@ -3145,8 +3145,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder,
> >   		 * load.
> >   		 */
> >   		DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n",
> > -			      pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp);
> > -		dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp;
> > +			      pipe_config->pipe_bpp, i915->vbt.edp.bpp);
> > +		i915->vbt.edp.bpp = pipe_config->pipe_bpp;
> >   	}
> >   }
> > @@ -3215,16 +3215,16 @@ static void chv_post_disable_dp(struct intel_encoder *encoder,
> >   				const struct intel_crtc_state *old_crtc_state,
> >   				const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	intel_dp_link_down(encoder, old_crtc_state);
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Assert data lane reset */
> >   	chv_data_lane_soft_reset(encoder, old_crtc_state, true);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   static void
> > @@ -3232,7 +3232,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> >   			 u32 *DP,
> >   			 u8 dp_train_pat)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum port port = intel_dig_port->base.port;
> >   	u8 train_pat_mask = drm_dp_training_pattern_mask(intel_dp->dpcd);
> > @@ -3241,7 +3241,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> >   		DRM_DEBUG_KMS("Using DP training pattern TPS%d\n",
> >   			      dp_train_pat & train_pat_mask);
> > -	if (HAS_DDI(dev_priv)) {
> > +	if (HAS_DDI(i915)) {
> >   		u32 temp = I915_READ(DP_TP_CTL(port));
> >   		if (dp_train_pat & DP_LINK_SCRAMBLING_DISABLE)
> > @@ -3270,8 +3270,8 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> >   		}
> >   		I915_WRITE(DP_TP_CTL(port), temp);
> > -	} else if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) ||
> > -		   (HAS_PCH_CPT(dev_priv) && port != PORT_A)) {
> > +	} else if ((IS_IVYBRIDGE(i915) && port == PORT_A) ||
> > +		   (HAS_PCH_CPT(i915) && port != PORT_A)) {
> >   		*DP &= ~DP_LINK_TRAIN_MASK_CPT;
> >   		switch (dp_train_pat & DP_TRAINING_PATTERN_MASK) {
> > @@ -3314,7 +3314,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp,
> >   static void intel_dp_enable_port(struct intel_dp *intel_dp,
> >   				 const struct intel_crtc_state *old_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	/* enable with pattern 1 (as per spec) */
> > @@ -3338,7 +3338,7 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> >   			    const struct intel_crtc_state *pipe_config,
> >   			    const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	u32 dp_reg = I915_READ(intel_dp->output_reg);
> > @@ -3349,7 +3349,7 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> >   		return;
> >   	with_pps_lock(intel_dp, wakeref) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   			vlv_init_panel_power_sequencer(encoder, pipe_config);
> >   		intel_dp_enable_port(intel_dp, pipe_config);
> > @@ -3359,13 +3359,13 @@ static void intel_enable_dp(struct intel_encoder *encoder,
> >   		edp_panel_vdd_off(intel_dp, true);
> >   	}
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		unsigned int lane_mask = 0x0;
> > -		if (IS_CHERRYVIEW(dev_priv))
> > +		if (IS_CHERRYVIEW(i915))
> >   			lane_mask = intel_dp_unused_lane_mask(pipe_config->lane_count);
> > -		vlv_wait_port_ready(dev_priv, dp_to_dig_port(intel_dp),
> > +		vlv_wait_port_ready(i915, dp_to_dig_port(intel_dp),
> >   				    lane_mask);
> >   	}
> > @@ -3412,7 +3412,7 @@ static void g4x_pre_enable_dp(struct intel_encoder *encoder,
> >   static void vlv_detach_power_sequencer(struct intel_dp *intel_dp)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> >   	enum pipe pipe = intel_dp->pps_pipe;
> >   	i915_reg_t pp_on_reg = PP_ON_DELAYS(pipe);
> > @@ -3440,14 +3440,14 @@ static void vlv_detach_power_sequencer(struct intel_dp *intel_dp)
> >   	intel_dp->pps_pipe = INVALID_PIPE;
> >   }
> > -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> > +static void vlv_steal_power_sequencer(struct drm_i915_private *i915,
> >   				      enum pipe pipe)
> >   {
> >   	struct intel_encoder *encoder;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> > -	for_each_intel_dp(&dev_priv->drm, encoder) {
> > +	for_each_intel_dp(&i915->drm, encoder) {
> >   		struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   		enum port port = encoder->port;
> > @@ -3469,11 +3469,11 @@ static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv,
> >   static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> >   					   const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	WARN_ON(intel_dp->active_pipe != INVALID_PIPE);
> > @@ -3491,7 +3491,7 @@ static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder,
> >   	 * We may be stealing the power
> >   	 * sequencer from another port.
> >   	 */
> > -	vlv_steal_power_sequencer(dev_priv, crtc->pipe);
> > +	vlv_steal_power_sequencer(i915, crtc->pipe);
> >   	intel_dp->active_pipe = crtc->pipe;
> > @@ -3570,17 +3570,17 @@ intel_dp_get_link_status(struct intel_dp *intel_dp, u8 link_status[DP_LINK_STATU
> >   u8
> >   intel_dp_voltage_max(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> >   	enum port port = encoder->port;
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		return intel_ddi_dp_voltage_max(encoder);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		return DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> > -	else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A)
> > +	else if (IS_IVYBRIDGE(i915) && port == PORT_A)
> >   		return DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> > -	else if (HAS_PCH_CPT(dev_priv) && port != PORT_A)
> > +	else if (HAS_PCH_CPT(i915) && port != PORT_A)
> >   		return DP_TRAIN_VOLTAGE_SWING_LEVEL_3;
> >   	else
> >   		return DP_TRAIN_VOLTAGE_SWING_LEVEL_2;
> > @@ -3589,13 +3589,13 @@ intel_dp_voltage_max(struct intel_dp *intel_dp)
> >   u8
> >   intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> >   	enum port port = encoder->port;
> > -	if (HAS_DDI(dev_priv)) {
> > +	if (HAS_DDI(i915)) {
> >   		return intel_ddi_dp_pre_emphasis_max(encoder, voltage_swing);
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) {
> >   		case DP_TRAIN_VOLTAGE_SWING_LEVEL_0:
> >   			return DP_TRAIN_PRE_EMPH_LEVEL_3;
> > @@ -3607,7 +3607,7 @@ intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing)
> >   		default:
> >   			return DP_TRAIN_PRE_EMPH_LEVEL_0;
> >   		}
> > -	} else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> > +	} else if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> >   		switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) {
> >   		case DP_TRAIN_VOLTAGE_SWING_LEVEL_0:
> >   			return DP_TRAIN_PRE_EMPH_LEVEL_2;
> > @@ -3901,25 +3901,25 @@ ivb_cpu_edp_signal_levels(u8 train_set)
> >   void
> >   intel_dp_set_signal_levels(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum port port = intel_dig_port->base.port;
> >   	u32 signal_levels, mask = 0;
> >   	u8 train_set = intel_dp->train_set[0];
> > -	if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +	if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10) {
> >   		signal_levels = bxt_signal_levels(intel_dp);
> > -	} else if (HAS_DDI(dev_priv)) {
> > +	} else if (HAS_DDI(i915)) {
> >   		signal_levels = ddi_signal_levels(intel_dp);
> >   		mask = DDI_BUF_EMP_MASK;
> > -	} else if (IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_CHERRYVIEW(i915)) {
> >   		signal_levels = chv_signal_levels(intel_dp);
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		signal_levels = vlv_signal_levels(intel_dp);
> > -	} else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) {
> > +	} else if (IS_IVYBRIDGE(i915) && port == PORT_A) {
> >   		signal_levels = ivb_cpu_edp_signal_levels(train_set);
> >   		mask = EDP_LINK_TRAIN_VOL_EMP_MASK_IVB;
> > -	} else if (IS_GEN(dev_priv, 6) && port == PORT_A) {
> > +	} else if (IS_GEN(i915, 6) && port == PORT_A) {
> >   		signal_levels = snb_cpu_edp_signal_levels(train_set);
> >   		mask = EDP_LINK_TRAIN_VOL_EMP_MASK_SNB;
> >   	} else {
> > @@ -3947,7 +3947,7 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp,
> >   				       u8 dp_train_pat)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(intel_dig_port->base.base.dev);
> >   	_intel_dp_set_link_train(intel_dp, &intel_dp->DP, dp_train_pat);
> > @@ -3958,12 +3958,12 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp,
> >   void intel_dp_set_idle_link_train(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	enum port port = intel_dig_port->base.port;
> >   	u32 val;
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		return;
> >   	val = I915_READ(DP_TP_CTL(port));
> > @@ -3981,7 +3981,7 @@ void intel_dp_set_idle_link_train(struct intel_dp *intel_dp)
> >   	if (port == PORT_A)
> >   		return;
> > -	if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port),
> > +	if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port),
> >   				    DP_TP_STATUS_IDLE_DONE,
> >   				    DP_TP_STATUS_IDLE_DONE,
> >   				    1))
> > @@ -3992,7 +3992,7 @@ static void
> >   intel_dp_link_down(struct intel_encoder *encoder,
> >   		   const struct intel_crtc_state *old_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	enum port port = encoder->port;
> > @@ -4003,8 +4003,8 @@ intel_dp_link_down(struct intel_encoder *encoder,
> >   	DRM_DEBUG_KMS("\n");
> > -	if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) ||
> > -	    (HAS_PCH_CPT(dev_priv) && port != PORT_A)) {
> > +	if ((IS_IVYBRIDGE(i915) && port == PORT_A) ||
> > +	    (HAS_PCH_CPT(i915) && port != PORT_A)) {
> >   		DP &= ~DP_LINK_TRAIN_MASK_CPT;
> >   		DP |= DP_LINK_TRAIN_PAT_IDLE_CPT;
> >   	} else {
> > @@ -4023,13 +4023,13 @@ intel_dp_link_down(struct intel_encoder *encoder,
> >   	 * to transcoder A after disabling it to allow the
> >   	 * matching HDMI port to be enabled on transcoder A.
> >   	 */
> > -	if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B && port != PORT_A) {
> > +	if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B && port != PORT_A) {
> >   		/*
> >   		 * We get CPU/PCH FIFO underruns on the other pipe when
> >   		 * doing the workaround. Sweep them under the rug.
> >   		 */
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> >   		/* always enable with pattern 1 (as per spec) */
> >   		DP &= ~(DP_PIPE_SEL_MASK | DP_LINK_TRAIN_MASK);
> > @@ -4042,16 +4042,16 @@ intel_dp_link_down(struct intel_encoder *encoder,
> >   		I915_WRITE(intel_dp->output_reg, DP);
> >   		POSTING_READ(intel_dp->output_reg);
> > -		intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > +		intel_wait_for_vblank_if_active(i915, PIPE_A);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> >   	}
> >   	msleep(intel_dp->panel_power_down_delay);
> >   	intel_dp->DP = DP;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		intel_wakeref_t wakeref;
> >   		with_pps_lock(intel_dp, wakeref)
> > @@ -4156,7 +4156,7 @@ static void intel_dp_get_dsc_sink_cap(struct intel_dp *intel_dp)
> >   static bool
> >   intel_edp_init_dpcd(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(dp_to_dig_port(intel_dp)->base.base.dev);
> >   	/* this function is meant to be called only once */
> > @@ -4169,7 +4169,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp)
> >   			 drm_dp_is_branch(intel_dp->dpcd));
> >   	if (intel_dp->dpcd[DP_DPCD_REV] >= 0x11)
> > -		dev_priv->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] &
> > +		i915->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] &
> >   			DP_NO_AUX_HANDSHAKE_LINK_TRAINING;
> >   	/*
> > @@ -4230,7 +4230,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp)
> >   	intel_dp_set_common_rates(intel_dp);
> >   	/* Read the eDP DSC DPCD registers */
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		intel_dp_get_dsc_sink_cap(intel_dp);
> >   	return true;
> > @@ -4804,7 +4804,7 @@ intel_dp_needs_link_retrain(struct intel_dp *intel_dp)
> >   int intel_dp_retrain_link(struct intel_encoder *encoder,
> >   			  struct drm_modeset_acquire_ctx *ctx)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_connector *connector = intel_dp->attached_connector;
> >   	struct drm_connector_state *conn_state;
> > @@ -4817,7 +4817,7 @@ int intel_dp_retrain_link(struct intel_encoder *encoder,
> >   	if (!connector || connector->base.status != connector_status_connected)
> >   		return 0;
> > -	ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex,
> > +	ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex,
> >   			       ctx);
> >   	if (ret)
> >   		return ret;
> > @@ -4847,20 +4847,20 @@ int intel_dp_retrain_link(struct intel_encoder *encoder,
> >   		return 0;
> >   	/* Suppress underruns caused by re-training */
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false);
> >   	if (crtc_state->has_pch_encoder)
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv,
> > +		intel_set_pch_fifo_underrun_reporting(i915,
> >   						      intel_crtc_pch_transcoder(crtc), false);
> >   	intel_dp_start_link_train(intel_dp);
> >   	intel_dp_stop_link_train(intel_dp);
> >   	/* Keep underrun reporting disabled until things are stable */
> > -	intel_wait_for_vblank(dev_priv, crtc->pipe);
> > +	intel_wait_for_vblank(i915, crtc->pipe);
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true);
> >   	if (crtc_state->has_pch_encoder)
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv,
> > +		intel_set_pch_fifo_underrun_reporting(i915,
> >   						      intel_crtc_pch_transcoder(crtc), true);
> >   	return 0;
> > @@ -4946,7 +4946,7 @@ static void intel_dp_check_service_irq(struct intel_dp *intel_dp)
> >   static bool
> >   intel_dp_short_pulse(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u8 old_sink_count = intel_dp->sink_count;
> >   	bool ret;
> > @@ -4983,7 +4983,7 @@ intel_dp_short_pulse(struct intel_dp *intel_dp)
> >   	if (intel_dp->compliance.test_type == DP_TEST_LINK_TRAINING) {
> >   		DRM_DEBUG_KMS("Link Training Compliance Test requested\n");
> >   		/* Send a Hotplug Uevent to userspace to start modeset */
> > -		drm_kms_helper_hotplug_event(&dev_priv->drm);
> > +		drm_kms_helper_hotplug_event(&i915->drm);
> >   	}
> >   	return true;
> > @@ -5052,7 +5052,7 @@ edp_detect(struct intel_dp *intel_dp)
> >   static bool ibx_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5075,7 +5075,7 @@ static bool ibx_digital_port_connected(struct intel_encoder *encoder)
> >   static bool cpt_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5098,7 +5098,7 @@ static bool cpt_digital_port_connected(struct intel_encoder *encoder)
> >   static bool spt_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5117,7 +5117,7 @@ static bool spt_digital_port_connected(struct intel_encoder *encoder)
> >   static bool g4x_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5140,7 +5140,7 @@ static bool g4x_digital_port_connected(struct intel_encoder *encoder)
> >   static bool gm45_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5163,7 +5163,7 @@ static bool gm45_digital_port_connected(struct intel_encoder *encoder)
> >   static bool ilk_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	if (encoder->hpd_pin == HPD_PORT_A)
> >   		return I915_READ(DEISR) & DE_DP_A_HOTPLUG;
> > @@ -5173,7 +5173,7 @@ static bool ilk_digital_port_connected(struct intel_encoder *encoder)
> >   static bool snb_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	if (encoder->hpd_pin == HPD_PORT_A)
> >   		return I915_READ(DEISR) & DE_DP_A_HOTPLUG;
> > @@ -5183,7 +5183,7 @@ static bool snb_digital_port_connected(struct intel_encoder *encoder)
> >   static bool ivb_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	if (encoder->hpd_pin == HPD_PORT_A)
> >   		return I915_READ(DEISR) & DE_DP_A_HOTPLUG_IVB;
> > @@ -5193,7 +5193,7 @@ static bool ivb_digital_port_connected(struct intel_encoder *encoder)
> >   static bool bdw_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	if (encoder->hpd_pin == HPD_PORT_A)
> >   		return I915_READ(GEN8_DE_PORT_ISR) & GEN8_PORT_DP_A_HOTPLUG;
> > @@ -5203,7 +5203,7 @@ static bool bdw_digital_port_connected(struct intel_encoder *encoder)
> >   static bool bxt_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 bit;
> >   	switch (encoder->hpd_pin) {
> > @@ -5224,7 +5224,7 @@ static bool bxt_digital_port_connected(struct intel_encoder *encoder)
> >   	return I915_READ(GEN8_DE_PORT_ISR) & bit;
> >   }
> > -static bool icl_combo_port_connected(struct drm_i915_private *dev_priv,
> > +static bool icl_combo_port_connected(struct drm_i915_private *i915,
> >   				     struct intel_digital_port *intel_dig_port)
> >   {
> >   	enum port port = intel_dig_port->base.port;
> > @@ -5247,7 +5247,7 @@ static const char *tc_type_name(enum tc_port_type type)
> >   	return names[type];
> >   }
> > -static void icl_update_tc_port_type(struct drm_i915_private *dev_priv,
> > +static void icl_update_tc_port_type(struct drm_i915_private *i915,
> >   				    struct intel_digital_port *intel_dig_port,
> >   				    bool is_legacy, bool is_typec, bool is_tbt)
> >   {
> > @@ -5295,10 +5295,10 @@ static void icl_update_tc_port_type(struct drm_i915_private *dev_priv,
> >    * will require a lot of coordination with user space and thorough testing for
> >    * the extra possible cases.
> >    */
> > -static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> > +static bool icl_tc_phy_connect(struct drm_i915_private *i915,
> >   			       struct intel_digital_port *dig_port)
> >   {
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
> >   	u32 val;
> >   	if (dig_port->tc_type != TC_PORT_LEGACY &&
> > @@ -5329,7 +5329,7 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> >   	if (dig_port->tc_type == TC_PORT_TYPEC &&
> >   	    !(I915_READ(PORT_TX_DFLEXDPSP) & TC_LIVE_STATE_TC(tc_port))) {
> >   		DRM_DEBUG_KMS("TC PHY %d sudden disconnect.\n", tc_port);
> > -		icl_tc_phy_disconnect(dev_priv, dig_port);
> > +		icl_tc_phy_disconnect(i915, dig_port);
> >   		return false;
> >   	}
> > @@ -5340,10 +5340,10 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv,
> >    * See the comment at the connect function. This implements the Disconnect
> >    * Flow.
> >    */
> > -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> > +void icl_tc_phy_disconnect(struct drm_i915_private *i915,
> >   			   struct intel_digital_port *dig_port)
> >   {
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port);
> >   	if (dig_port->tc_type == TC_PORT_UNKNOWN)
> >   		return;
> > @@ -5378,11 +5378,11 @@ void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> >    * connected ports are usable, and avoids exposing to the users objects they
> >    * can't really use.
> >    */
> > -static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
> > +static bool icl_tc_port_connected(struct drm_i915_private *i915,
> >   				  struct intel_digital_port *intel_dig_port)
> >   {
> >   	enum port port = intel_dig_port->base.port;
> > -	enum tc_port tc_port = intel_port_to_tc(dev_priv, port);
> > +	enum tc_port tc_port = intel_port_to_tc(i915, port);
> >   	bool is_legacy, is_typec, is_tbt;
> >   	u32 dpsp;
> > @@ -5407,15 +5407,15 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
> >   	is_tbt = dpsp & TC_LIVE_STATE_TBT(tc_port);
> >   	if (!is_legacy && !is_typec && !is_tbt) {
> > -		icl_tc_phy_disconnect(dev_priv, intel_dig_port);
> > +		icl_tc_phy_disconnect(i915, intel_dig_port);
> >   		return false;
> >   	}
> > -	icl_update_tc_port_type(dev_priv, intel_dig_port, is_legacy, is_typec,
> > +	icl_update_tc_port_type(i915, intel_dig_port, is_legacy, is_typec,
> >   				is_tbt);
> > -	if (!icl_tc_phy_connect(dev_priv, intel_dig_port))
> > +	if (!icl_tc_phy_connect(i915, intel_dig_port))
> >   		return false;
> >   	return true;
> > @@ -5423,13 +5423,13 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv,
> >   static bool icl_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> > -	if (intel_port_is_combophy(dev_priv, encoder->port))
> > -		return icl_combo_port_connected(dev_priv, dig_port);
> > -	else if (intel_port_is_tc(dev_priv, encoder->port))
> > -		return icl_tc_port_connected(dev_priv, dig_port);
> > +	if (intel_port_is_combophy(i915, encoder->port))
> > +		return icl_combo_port_connected(i915, dig_port);
> > +	else if (intel_port_is_tc(i915, encoder->port))
> > +		return icl_tc_port_connected(i915, dig_port);
> >   	else
> >   		MISSING_CASE(encoder->hpd_pin);
> > @@ -5449,41 +5449,41 @@ static bool icl_digital_port_connected(struct intel_encoder *encoder)
> >    */
> >   static bool __intel_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > -	if (HAS_GMCH(dev_priv)) {
> > -		if (IS_GM45(dev_priv))
> > +	if (HAS_GMCH(i915)) {
> > +		if (IS_GM45(i915))
> >   			return gm45_digital_port_connected(encoder);
> >   		else
> >   			return g4x_digital_port_connected(encoder);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return icl_digital_port_connected(encoder);
> > -	else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv))
> > +	else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915))
> >   		return spt_digital_port_connected(encoder);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		return bxt_digital_port_connected(encoder);
> > -	else if (IS_GEN(dev_priv, 8))
> > +	else if (IS_GEN(i915, 8))
> >   		return bdw_digital_port_connected(encoder);
> > -	else if (IS_GEN(dev_priv, 7))
> > +	else if (IS_GEN(i915, 7))
> >   		return ivb_digital_port_connected(encoder);
> > -	else if (IS_GEN(dev_priv, 6))
> > +	else if (IS_GEN(i915, 6))
> >   		return snb_digital_port_connected(encoder);
> > -	else if (IS_GEN(dev_priv, 5))
> > +	else if (IS_GEN(i915, 5))
> >   		return ilk_digital_port_connected(encoder);
> > -	MISSING_CASE(INTEL_GEN(dev_priv));
> > +	MISSING_CASE(INTEL_GEN(i915));
> >   	return false;
> >   }
> >   bool intel_digital_port_connected(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	bool is_connected = false;
> >   	intel_wakeref_t wakeref;
> > -	with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> > +	with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref)
> >   		is_connected = __intel_digital_port_connected(encoder);
> >   	return is_connected;
> > @@ -5537,7 +5537,7 @@ intel_dp_detect(struct drm_connector *connector,
> >   		struct drm_modeset_acquire_ctx *ctx,
> >   		bool force)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_dp *intel_dp = intel_attached_dp(connector);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	struct intel_encoder *encoder = &dig_port->base;
> > @@ -5545,7 +5545,7 @@ intel_dp_detect(struct drm_connector *connector,
> >   	DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> >   		      connector->base.id, connector->name);
> > -	WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> > +	WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
> >   	/* Can't disconnect eDP */
> >   	if (intel_dp_is_edp(intel_dp))
> > @@ -5584,7 +5584,7 @@ intel_dp_detect(struct drm_connector *connector,
> >   	intel_dp_print_rates(intel_dp);
> >   	/* Read DP Sink DSC Cap DPCD regs for DP v1.4 */
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		intel_dp_get_dsc_sink_cap(intel_dp);
> >   	drm_dp_read_desc(&intel_dp->aux, &intel_dp->desc,
> > @@ -5642,7 +5642,7 @@ intel_dp_force(struct drm_connector *connector)
> >   	struct intel_dp *intel_dp = intel_attached_dp(connector);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> >   	struct intel_encoder *intel_encoder = &dig_port->base;
> > -	struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> >   	enum intel_display_power_domain aux_domain =
> >   		intel_aux_power_domain(dig_port);
> >   	intel_wakeref_t wakeref;
> > @@ -5654,11 +5654,11 @@ intel_dp_force(struct drm_connector *connector)
> >   	if (connector->status != connector_status_connected)
> >   		return;
> > -	wakeref = intel_display_power_get(dev_priv, aux_domain);
> > +	wakeref = intel_display_power_get(i915, aux_domain);
> >   	intel_dp_set_edid(intel_dp);
> > -	intel_display_power_put(dev_priv, aux_domain, wakeref);
> > +	intel_display_power_put(i915, aux_domain, wakeref);
> >   }
> >   static int intel_dp_get_modes(struct drm_connector *connector)
> > @@ -6362,10 +6362,10 @@ static const struct intel_hdcp_shim intel_dp_hdcp_shim = {
> >   static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	if (!edp_have_panel_vdd(intel_dp))
> >   		return;
> > @@ -6377,18 +6377,18 @@ static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp)
> >   	 * indefinitely.
> >   	 */
> >   	DRM_DEBUG_KMS("VDD left on by BIOS, adjusting state tracking\n");
> > -	intel_display_power_get(dev_priv, intel_aux_power_domain(dig_port));
> > +	intel_display_power_get(i915, intel_aux_power_domain(dig_port));
> >   	edp_panel_vdd_schedule_off(intel_dp);
> >   }
> >   static enum pipe vlv_active_pipe(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base;
> >   	enum pipe pipe;
> > -	if (intel_dp_port_enabled(dev_priv, intel_dp->output_reg,
> > +	if (intel_dp_port_enabled(i915, intel_dp->output_reg,
> >   				  encoder->port, &pipe))
> >   		return pipe;
> > @@ -6397,12 +6397,12 @@ static enum pipe vlv_active_pipe(struct intel_dp *intel_dp)
> >   void intel_dp_encoder_reset(struct drm_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(encoder);
> >   	struct intel_lspcon *lspcon = dp_to_lspcon(intel_dp);
> >   	intel_wakeref_t wakeref;
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		intel_dp->DP = I915_READ(intel_dp->output_reg);
> >   	if (lspcon->active)
> > @@ -6410,12 +6410,12 @@ void intel_dp_encoder_reset(struct drm_encoder *encoder)
> >   	intel_dp->reset_link_params = true;
> > -	if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) &&
> > +	if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) &&
> >   	    !intel_dp_is_edp(intel_dp))
> >   		return;
> >   	with_pps_lock(intel_dp, wakeref) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   			intel_dp->active_pipe = vlv_active_pipe(intel_dp);
> >   		if (intel_dp_is_edp(intel_dp)) {
> > @@ -6508,41 +6508,41 @@ intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port, bool long_hpd)
> >   }
> >   /* check the VBT to see whether the eDP is on another port */
> > -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port)
> > +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port)
> >   {
> >   	/*
> >   	 * eDP not supported on g4x. so bail out early just
> >   	 * for a bit extra safety in case the VBT is bonkers.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) < 5)
> > +	if (INTEL_GEN(i915) < 5)
> >   		return false;
> > -	if (INTEL_GEN(dev_priv) < 9 && port == PORT_A)
> > +	if (INTEL_GEN(i915) < 9 && port == PORT_A)
> >   		return true;
> > -	return intel_bios_is_port_edp(dev_priv, port);
> > +	return intel_bios_is_port_edp(i915, port);
> >   }
> >   static void
> >   intel_dp_add_properties(struct intel_dp *intel_dp, struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	enum port port = dp_to_dig_port(intel_dp)->base.port;
> > -	if (!IS_G4X(dev_priv) && port != PORT_A)
> > +	if (!IS_G4X(i915) && port != PORT_A)
> >   		intel_attach_force_audio_property(connector);
> >   	intel_attach_broadcast_rgb_property(connector);
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		drm_connector_attach_max_bpc_property(connector, 6, 10);
> > -	else if (INTEL_GEN(dev_priv) >= 5)
> > +	else if (INTEL_GEN(i915) >= 5)
> >   		drm_connector_attach_max_bpc_property(connector, 6, 12);
> >   	if (intel_dp_is_edp(intel_dp)) {
> >   		u32 allowed_scalers;
> >   		allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN);
> > -		if (!HAS_GMCH(dev_priv))
> > +		if (!HAS_GMCH(i915))
> >   			allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER);
> >   		drm_connector_attach_scaling_mode_property(connector, allowed_scalers);
> > @@ -6562,7 +6562,7 @@ static void intel_dp_init_panel_power_timestamps(struct intel_dp *intel_dp)
> >   static void
> >   intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 pp_on, pp_off, pp_ctl;
> >   	struct pps_registers regs;
> > @@ -6571,7 +6571,7 @@ intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq)
> >   	pp_ctl = ironlake_get_pp_control(intel_dp);
> >   	/* Ensure PPS is unlocked */
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		I915_WRITE(regs.pp_ctrl, pp_ctl);
> >   	pp_on = I915_READ(regs.pp_on);
> > @@ -6621,11 +6621,11 @@ intel_pps_verify_state(struct intel_dp *intel_dp)
> >   static void
> >   intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct edp_power_seq cur, vbt, spec,
> >   		*final = &intel_dp->pps_delays;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	/* already initialized? */
> >   	if (final->t11_t12 != 0)
> > @@ -6635,13 +6635,13 @@ intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp)
> >   	intel_pps_dump_state("cur", &cur);
> > -	vbt = dev_priv->vbt.edp.pps;
> > +	vbt = i915->vbt.edp.pps;
> >   	/* On Toshiba Satellite P50-C-18C system the VBT T12 delay
> >   	 * of 500ms appears to be too short. Ocassionally the panel
> >   	 * just fails to power back on. Increasing the delay to 800ms
> >   	 * seems sufficient to avoid this problem.
> >   	 */
> > -	if (dev_priv->quirks & QUIRK_INCREASE_T12_DELAY) {
> > +	if (i915->quirks & QUIRK_INCREASE_T12_DELAY) {
> >   		vbt.t11_t12 = max_t(u16, vbt.t11_t12, 1300 * 10);
> >   		DRM_DEBUG_KMS("Increasing T12 panel delay as per the quirk to %d\n",
> >   			      vbt.t11_t12);
> > @@ -6714,14 +6714,14 @@ static void
> >   intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
> >   					      bool force_disable_vdd)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 pp_on, pp_off, port_sel = 0;
> > -	int div = dev_priv->rawclk_freq / 1000;
> > +	int div = i915->rawclk_freq / 1000;
> >   	struct pps_registers regs;
> >   	enum port port = dp_to_dig_port(intel_dp)->base.port;
> >   	const struct edp_power_seq *seq = &intel_dp->pps_delays;
> > -	lockdep_assert_held(&dev_priv->pps_mutex);
> > +	lockdep_assert_held(&i915->pps_mutex);
> >   	intel_pps_get_registers(intel_dp, &regs);
> > @@ -6757,9 +6757,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
> >   	/* Haswell doesn't have any port selection bits for the panel
> >   	 * power sequencer any more. */
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		port_sel = PANEL_PORT_SELECT_VLV(port);
> > -	} else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) {
> > +	} else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) {
> >   		switch (port) {
> >   		case PORT_A:
> >   			port_sel = PANEL_PORT_SELECT_DPA;
> > @@ -6807,9 +6807,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp,
> >   static void intel_dp_pps_init(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		vlv_initial_power_sequencer_setup(intel_dp);
> >   	} else {
> >   		intel_dp_init_panel_power_sequencer(intel_dp);
> > @@ -6819,7 +6819,7 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp)
> >   /**
> >    * intel_dp_set_drrs_state - program registers for RR switch to take effect
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @crtc_state: a pointer to the active intel_crtc_state
> >    * @refresh_rate: RR to be programmed
> >    *
> > @@ -6828,15 +6828,15 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp)
> >    * supported by the panel or to any other RR based on media playback (in
> >    * this case, RR value needs to be passed from user space).
> >    *
> > - * The caller of this function needs to take a lock on dev_priv->drrs.
> > + * The caller of this function needs to take a lock on i915->drrs.
> >    */
> > -static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> > +static void intel_dp_set_drrs_state(struct drm_i915_private *i915,
> >   				    const struct intel_crtc_state *crtc_state,
> >   				    int refresh_rate)
> >   {
> >   	struct intel_encoder *encoder;
> >   	struct intel_digital_port *dig_port = NULL;
> > -	struct intel_dp *intel_dp = dev_priv->drrs.dp;
> > +	struct intel_dp *intel_dp = i915->drrs.dp;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum drrs_refresh_rate_type index = DRRS_HIGH_RR;
> > @@ -6858,7 +6858,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	if (dev_priv->drrs.type < SEAMLESS_DRRS_SUPPORT) {
> > +	if (i915->drrs.type < SEAMLESS_DRRS_SUPPORT) {
> >   		DRM_DEBUG_KMS("Only Seamless DRRS supported.\n");
> >   		return;
> >   	}
> > @@ -6867,7 +6867,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   			refresh_rate)
> >   		index = DRRS_LOW_RR;
> > -	if (index == dev_priv->drrs.refresh_rate_type) {
> > +	if (index == i915->drrs.refresh_rate_type) {
> >   		DRM_DEBUG_KMS(
> >   			"DRRS requested for previously set RR...ignoring\n");
> >   		return;
> > @@ -6878,7 +6878,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   		return;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 8 && !IS_CHERRYVIEW(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 8 && !IS_CHERRYVIEW(i915)) {
> >   		switch (index) {
> >   		case DRRS_HIGH_RR:
> >   			intel_dp_set_m_n(crtc_state, M1_N1);
> > @@ -6890,18 +6890,18 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   		default:
> >   			DRM_ERROR("Unsupported refreshrate type\n");
> >   		}
> > -	} else if (INTEL_GEN(dev_priv) > 6) {
> > +	} else if (INTEL_GEN(i915) > 6) {
> >   		i915_reg_t reg = PIPECONF(crtc_state->cpu_transcoder);
> >   		u32 val;
> >   		val = I915_READ(reg);
> >   		if (index > DRRS_HIGH_RR) {
> > -			if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +			if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   				val |= PIPECONF_EDP_RR_MODE_SWITCH_VLV;
> >   			else
> >   				val |= PIPECONF_EDP_RR_MODE_SWITCH;
> >   		} else {
> > -			if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +			if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   				val &= ~PIPECONF_EDP_RR_MODE_SWITCH_VLV;
> >   			else
> >   				val &= ~PIPECONF_EDP_RR_MODE_SWITCH;
> > @@ -6909,7 +6909,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   		I915_WRITE(reg, val);
> >   	}
> > -	dev_priv->drrs.refresh_rate_type = index;
> > +	i915->drrs.refresh_rate_type = index;
> >   	DRM_DEBUG_KMS("eDP Refresh Rate set to : %dHz\n", refresh_rate);
> >   }
> > @@ -6924,30 +6924,30 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv,
> >   void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> >   			   const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (!crtc_state->has_drrs) {
> >   		DRM_DEBUG_KMS("Panel doesn't support DRRS\n");
> >   		return;
> >   	}
> > -	if (dev_priv->psr.enabled) {
> > +	if (i915->psr.enabled) {
> >   		DRM_DEBUG_KMS("PSR enabled. Not enabling DRRS.\n");
> >   		return;
> >   	}
> > -	mutex_lock(&dev_priv->drrs.mutex);
> > -	if (dev_priv->drrs.dp) {
> > +	mutex_lock(&i915->drrs.mutex);
> > +	if (i915->drrs.dp) {
> >   		DRM_DEBUG_KMS("DRRS already enabled\n");
> >   		goto unlock;
> >   	}
> > -	dev_priv->drrs.busy_frontbuffer_bits = 0;
> > +	i915->drrs.busy_frontbuffer_bits = 0;
> > -	dev_priv->drrs.dp = intel_dp;
> > +	i915->drrs.dp = intel_dp;
> >   unlock:
> > -	mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_unlock(&i915->drrs.mutex);
> >   }
> >   /**
> > @@ -6959,36 +6959,36 @@ void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> >   void intel_edp_drrs_disable(struct intel_dp *intel_dp,
> >   			    const struct intel_crtc_state *old_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (!old_crtc_state->has_drrs)
> >   		return;
> > -	mutex_lock(&dev_priv->drrs.mutex);
> > -	if (!dev_priv->drrs.dp) {
> > -		mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_lock(&i915->drrs.mutex);
> > +	if (!i915->drrs.dp) {
> > +		mutex_unlock(&i915->drrs.mutex);
> >   		return;
> >   	}
> > -	if (dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> > -		intel_dp_set_drrs_state(dev_priv, old_crtc_state,
> > +	if (i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> > +		intel_dp_set_drrs_state(i915, old_crtc_state,
> >   			intel_dp->attached_connector->panel.fixed_mode->vrefresh);
> > -	dev_priv->drrs.dp = NULL;
> > -	mutex_unlock(&dev_priv->drrs.mutex);
> > +	i915->drrs.dp = NULL;
> > +	mutex_unlock(&i915->drrs.mutex);
> > -	cancel_delayed_work_sync(&dev_priv->drrs.work);
> > +	cancel_delayed_work_sync(&i915->drrs.work);
> >   }
> >   static void intel_edp_drrs_downclock_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv), drrs.work.work);
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915), drrs.work.work);
> >   	struct intel_dp *intel_dp;
> > -	mutex_lock(&dev_priv->drrs.mutex);
> > +	mutex_lock(&i915->drrs.mutex);
> > -	intel_dp = dev_priv->drrs.dp;
> > +	intel_dp = i915->drrs.dp;
> >   	if (!intel_dp)
> >   		goto unlock;
> > @@ -6998,23 +6998,23 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work)
> >   	 * recheck.
> >   	 */
> > -	if (dev_priv->drrs.busy_frontbuffer_bits)
> > +	if (i915->drrs.busy_frontbuffer_bits)
> >   		goto unlock;
> > -	if (dev_priv->drrs.refresh_rate_type != DRRS_LOW_RR) {
> > +	if (i915->drrs.refresh_rate_type != DRRS_LOW_RR) {
> >   		struct drm_crtc *crtc = dp_to_dig_port(intel_dp)->base.base.crtc;
> > -		intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> > +		intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> >   			intel_dp->attached_connector->panel.downclock_mode->vrefresh);
> >   	}
> >   unlock:
> > -	mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_unlock(&i915->drrs.mutex);
> >   }
> >   /**
> >    * intel_edp_drrs_invalidate - Disable Idleness DRRS
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    *
> >    * This function gets called everytime rendering on the given planes start.
> > @@ -7022,40 +7022,40 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work)
> >    *
> >    * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
> >    */
> > -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> > +void intel_edp_drrs_invalidate(struct drm_i915_private *i915,
> >   			       unsigned int frontbuffer_bits)
> >   {
> >   	struct drm_crtc *crtc;
> >   	enum pipe pipe;
> > -	if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED)
> > +	if (i915->drrs.type == DRRS_NOT_SUPPORTED)
> >   		return;
> > -	cancel_delayed_work(&dev_priv->drrs.work);
> > +	cancel_delayed_work(&i915->drrs.work);
> > -	mutex_lock(&dev_priv->drrs.mutex);
> > -	if (!dev_priv->drrs.dp) {
> > -		mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_lock(&i915->drrs.mutex);
> > +	if (!i915->drrs.dp) {
> > +		mutex_unlock(&i915->drrs.mutex);
> >   		return;
> >   	}
> > -	crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc;
> > +	crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc;
> >   	pipe = to_intel_crtc(crtc)->pipe;
> >   	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe);
> > -	dev_priv->drrs.busy_frontbuffer_bits |= frontbuffer_bits;
> > +	i915->drrs.busy_frontbuffer_bits |= frontbuffer_bits;
> >   	/* invalidate means busy screen hence upclock */
> > -	if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> > -		intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> > -			dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> > +	if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> > +		intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> > +			i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> > -	mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_unlock(&i915->drrs.mutex);
> >   }
> >   /**
> >    * intel_edp_drrs_flush - Restart Idleness DRRS
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    *
> >    * This function gets called every time rendering on the given planes has
> > @@ -7065,42 +7065,42 @@ void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> >    *
> >    * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits.
> >    */
> > -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv,
> > +void intel_edp_drrs_flush(struct drm_i915_private *i915,
> >   			  unsigned int frontbuffer_bits)
> >   {
> >   	struct drm_crtc *crtc;
> >   	enum pipe pipe;
> > -	if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED)
> > +	if (i915->drrs.type == DRRS_NOT_SUPPORTED)
> >   		return;
> > -	cancel_delayed_work(&dev_priv->drrs.work);
> > +	cancel_delayed_work(&i915->drrs.work);
> > -	mutex_lock(&dev_priv->drrs.mutex);
> > -	if (!dev_priv->drrs.dp) {
> > -		mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_lock(&i915->drrs.mutex);
> > +	if (!i915->drrs.dp) {
> > +		mutex_unlock(&i915->drrs.mutex);
> >   		return;
> >   	}
> > -	crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc;
> > +	crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc;
> >   	pipe = to_intel_crtc(crtc)->pipe;
> >   	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe);
> > -	dev_priv->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits;
> > +	i915->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits;
> >   	/* flush means busy screen hence upclock */
> > -	if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR)
> > -		intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config,
> > -				dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> > +	if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR)
> > +		intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config,
> > +				i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh);
> >   	/*
> >   	 * flush also means no more activity hence schedule downclock, if all
> >   	 * other fbs are quiescent too
> >   	 */
> > -	if (!dev_priv->drrs.busy_frontbuffer_bits)
> > -		schedule_delayed_work(&dev_priv->drrs.work,
> > +	if (!i915->drrs.busy_frontbuffer_bits)
> > +		schedule_delayed_work(&i915->drrs.work,
> >   				msecs_to_jiffies(1000));
> > -	mutex_unlock(&dev_priv->drrs.mutex);
> > +	mutex_unlock(&i915->drrs.mutex);
> >   }
> >   /**
> > @@ -7157,18 +7157,18 @@ static struct drm_display_mode *
> >   intel_dp_drrs_init(struct intel_connector *connector,
> >   		   struct drm_display_mode *fixed_mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct drm_display_mode *downclock_mode = NULL;
> > -	INIT_DELAYED_WORK(&dev_priv->drrs.work, intel_edp_drrs_downclock_work);
> > -	mutex_init(&dev_priv->drrs.mutex);
> > +	INIT_DELAYED_WORK(&i915->drrs.work, intel_edp_drrs_downclock_work);
> > +	mutex_init(&i915->drrs.mutex);
> > -	if (INTEL_GEN(dev_priv) <= 6) {
> > +	if (INTEL_GEN(i915) <= 6) {
> >   		DRM_DEBUG_KMS("DRRS supported for Gen7 and above\n");
> >   		return NULL;
> >   	}
> > -	if (dev_priv->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) {
> > +	if (i915->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) {
> >   		DRM_DEBUG_KMS("VBT doesn't support DRRS\n");
> >   		return NULL;
> >   	}
> > @@ -7179,9 +7179,9 @@ intel_dp_drrs_init(struct intel_connector *connector,
> >   		return NULL;
> >   	}
> > -	dev_priv->drrs.type = dev_priv->vbt.drrs_type;
> > +	i915->drrs.type = i915->vbt.drrs_type;
> > -	dev_priv->drrs.refresh_rate_type = DRRS_HIGH_RR;
> > +	i915->drrs.refresh_rate_type = DRRS_HIGH_RR;
> >   	DRM_DEBUG_KMS("seamless DRRS supported for eDP panel.\n");
> >   	return downclock_mode;
> >   }
> > @@ -7189,8 +7189,8 @@ intel_dp_drrs_init(struct intel_connector *connector,
> >   static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> >   				     struct intel_connector *intel_connector)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector = &intel_connector->base;
> >   	struct drm_display_mode *fixed_mode = NULL;
> >   	struct drm_display_mode *downclock_mode = NULL;
> > @@ -7210,8 +7210,8 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> >   	 * eDP and LVDS bail out early in this case to prevent interfering
> >   	 * with an already powered-on LVDS power sequencer.
> >   	 */
> > -	if (intel_get_lvds_encoder(dev_priv)) {
> > -		WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)));
> > +	if (intel_get_lvds_encoder(i915)) {
> > +		WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)));
> >   		DRM_INFO("LVDS was detected, not registering eDP\n");
> >   		return false;
> > @@ -7256,7 +7256,7 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp,
> >   		fixed_mode = intel_panel_vbt_fixed_mode(intel_connector);
> >   	mutex_unlock(&dev->mode_config.mutex);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		intel_dp->edp_notifier.notifier_call = edp_notify_handler;
> >   		register_reboot_notifier(&intel_dp->edp_notifier);
> > @@ -7330,7 +7330,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	struct intel_dp *intel_dp = &intel_dig_port->dp;
> >   	struct intel_encoder *intel_encoder = &intel_dig_port->base;
> >   	struct drm_device *dev = intel_encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum port port = intel_encoder->port;
> >   	int type;
> > @@ -7353,18 +7353,18 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	intel_dp->DP = I915_READ(intel_dp->output_reg);
> >   	intel_dp->attached_connector = intel_connector;
> > -	if (intel_dp_is_port_edp(dev_priv, port)) {
> > +	if (intel_dp_is_port_edp(i915, port)) {
> >   		/*
> >   		 * Currently we don't support eDP on TypeC ports, although in
> >   		 * theory it could work on TypeC legacy ports.
> >   		 */
> > -		WARN_ON(intel_port_is_tc(dev_priv, port));
> > +		WARN_ON(intel_port_is_tc(i915, port));
> >   		type = DRM_MODE_CONNECTOR_eDP;
> >   	} else {
> >   		type = DRM_MODE_CONNECTOR_DisplayPort;
> >   	}
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		intel_dp->active_pipe = vlv_active_pipe(intel_dp);
> >   	/*
> > @@ -7376,7 +7376,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   		intel_encoder->type = INTEL_OUTPUT_EDP;
> >   	/* eDP only on port B and/or C on vlv/chv */
> > -	if (WARN_ON((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if (WARN_ON((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   		    intel_dp_is_edp(intel_dp) &&
> >   		    port != PORT_B && port != PORT_C))
> >   		return false;
> > @@ -7388,26 +7388,26 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	drm_connector_init(dev, connector, &intel_dp_connector_funcs, type);
> >   	drm_connector_helper_add(connector, &intel_dp_connector_helper_funcs);
> > -	if (!HAS_GMCH(dev_priv))
> > +	if (!HAS_GMCH(i915))
> >   		connector->interlace_allowed = true;
> >   	connector->doublescan_allowed = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		connector->ycbcr_420_allowed = true;
> > -	intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port);
> > +	intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port);
> >   	intel_dp_aux_init(intel_dp);
> >   	intel_connector_attach_encoder(intel_connector, intel_encoder);
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		intel_connector->get_hw_state = intel_ddi_connector_get_hw_state;
> >   	else
> >   		intel_connector->get_hw_state = intel_connector_get_hw_state;
> >   	/* init MST on ports that can support it */
> > -	if (HAS_DP_MST(dev_priv) && !intel_dp_is_edp(intel_dp) &&
> > +	if (HAS_DP_MST(i915) && !intel_dp_is_edp(intel_dp) &&
> >   	    (port == PORT_B || port == PORT_C ||
> >   	     port == PORT_D || port == PORT_F))
> >   		intel_dp_mst_encoder_init(intel_dig_port,
> > @@ -7421,7 +7421,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	intel_dp_add_properties(intel_dp, connector);
> > -	if (is_hdcp_supported(dev_priv, port) && !intel_dp_is_edp(intel_dp)) {
> > +	if (is_hdcp_supported(i915, port) && !intel_dp_is_edp(intel_dp)) {
> >   		int ret = intel_hdcp_init(intel_connector, &intel_dp_hdcp_shim);
> >   		if (ret)
> >   			DRM_DEBUG_KMS("HDCP init failed, skipping.\n");
> > @@ -7431,7 +7431,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	 * 0xd.  Failure to do so will result in spurious interrupts being
> >   	 * generated on the port when a cable is not attached.
> >   	 */
> > -	if (IS_G45(dev_priv)) {
> > +	if (IS_G45(i915)) {
> >   		u32 temp = I915_READ(PEG_BAND_GAP_DATA);
> >   		I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd);
> >   	}
> > @@ -7444,7 +7444,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   	return false;
> >   }
> > -bool intel_dp_init(struct drm_i915_private *dev_priv,
> > +bool intel_dp_init(struct drm_i915_private *i915,
> >   		   i915_reg_t output_reg,
> >   		   enum port port)
> >   {
> > @@ -7464,7 +7464,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder = &intel_dig_port->base;
> >   	encoder = &intel_encoder->base;
> > -	if (drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> > +	if (drm_encoder_init(&i915->drm, &intel_encoder->base,
> >   			     &intel_dp_enc_funcs, DRM_MODE_ENCODER_TMDS,
> >   			     "DP %c", port_name(port)))
> >   		goto err_encoder_init;
> > @@ -7475,14 +7475,14 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder->get_config = intel_dp_get_config;
> >   	intel_encoder->update_pipe = intel_panel_update_backlight;
> >   	intel_encoder->suspend = intel_dp_encoder_suspend;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		intel_encoder->pre_pll_enable = chv_dp_pre_pll_enable;
> >   		intel_encoder->pre_enable = chv_pre_enable_dp;
> >   		intel_encoder->enable = vlv_enable_dp;
> >   		intel_encoder->disable = vlv_disable_dp;
> >   		intel_encoder->post_disable = chv_post_disable_dp;
> >   		intel_encoder->post_pll_disable = chv_dp_post_pll_disable;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		intel_encoder->pre_pll_enable = vlv_dp_pre_pll_enable;
> >   		intel_encoder->pre_enable = vlv_pre_enable_dp;
> >   		intel_encoder->enable = vlv_enable_dp;
> > @@ -7500,7 +7500,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder->type = INTEL_OUTPUT_DP;
> >   	intel_encoder->power_domain = intel_port_to_power_domain(port);
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		if (port == PORT_D)
> >   			intel_encoder->crtc_mask = 1 << 2;
> >   		else
> > @@ -7516,7 +7516,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> >   	if (port != PORT_A)
> >   		intel_infoframe_init(intel_dig_port);
> > -	intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> > +	intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
> >   	if (!intel_dp_init_connector(intel_dig_port, intel_connector))
> >   		goto err_init_connector;
> > @@ -7531,11 +7531,11 @@ bool intel_dp_init(struct drm_i915_private *dev_priv,
> >   	return false;
> >   }
> > -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv)
> > +void intel_dp_mst_suspend(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		struct intel_dp *intel_dp;
> >   		if (encoder->type != INTEL_OUTPUT_DDI)
> > @@ -7551,11 +7551,11 @@ void intel_dp_mst_suspend(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -void intel_dp_mst_resume(struct drm_i915_private *dev_priv)
> > +void intel_dp_mst_resume(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		struct intel_dp *intel_dp;
> >   		int ret;
> > diff --git a/drivers/gpu/drm/i915/intel_dp.h b/drivers/gpu/drm/i915/intel_dp.h
> > index da70b1a41c83..b6fecb185b23 100644
> > --- a/drivers/gpu/drm/i915/intel_dp.h
> > +++ b/drivers/gpu/drm/i915/intel_dp.h
> > @@ -35,10 +35,10 @@ void intel_dp_adjust_compliance_config(struct intel_dp *intel_dp,
> >   bool intel_dp_limited_color_range(const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state);
> >   int intel_dp_min_bpp(const struct intel_crtc_state *crtc_state);
> > -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_dp_port_enabled(struct drm_i915_private *i915,
> >   			   i915_reg_t dp_reg, enum port port,
> >   			   enum pipe *pipe);
> > -bool intel_dp_init(struct drm_i915_private *dev_priv, i915_reg_t output_reg,
> > +bool intel_dp_init(struct drm_i915_private *i915, i915_reg_t output_reg,
> >   		   enum port port);
> >   bool intel_dp_init_connector(struct intel_digital_port *intel_dig_port,
> >   			     struct intel_connector *intel_connector);
> > @@ -60,7 +60,7 @@ int intel_dp_compute_config(struct intel_encoder *encoder,
> >   			    struct intel_crtc_state *pipe_config,
> >   			    struct drm_connector_state *conn_state);
> >   bool intel_dp_is_edp(struct intel_dp *intel_dp);
> > -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port);
> > +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port);
> >   enum irqreturn intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port,
> >   				  bool long_hpd);
> >   void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state,
> > @@ -69,21 +69,21 @@ void intel_edp_backlight_off(const struct drm_connector_state *conn_state);
> >   void intel_edp_panel_vdd_on(struct intel_dp *intel_dp);
> >   void intel_edp_panel_on(struct intel_dp *intel_dp);
> >   void intel_edp_panel_off(struct intel_dp *intel_dp);
> > -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv);
> > -void intel_dp_mst_resume(struct drm_i915_private *dev_priv);
> > +void intel_dp_mst_suspend(struct drm_i915_private *i915);
> > +void intel_dp_mst_resume(struct drm_i915_private *i915);
> >   int intel_dp_max_link_rate(struct intel_dp *intel_dp);
> >   int intel_dp_max_lane_count(struct intel_dp *intel_dp);
> >   int intel_dp_rate_select(struct intel_dp *intel_dp, int rate);
> > -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv);
> > +void intel_power_sequencer_reset(struct drm_i915_private *i915);
> >   u32 intel_dp_pack_aux(const u8 *src, int src_bytes);
> >   void intel_edp_drrs_enable(struct intel_dp *intel_dp,
> >   			   const struct intel_crtc_state *crtc_state);
> >   void intel_edp_drrs_disable(struct intel_dp *intel_dp,
> >   			    const struct intel_crtc_state *crtc_state);
> > -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv,
> > +void intel_edp_drrs_invalidate(struct drm_i915_private *i915,
> >   			       unsigned int frontbuffer_bits);
> > -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv,
> > +void intel_edp_drrs_flush(struct drm_i915_private *i915,
> >   			  unsigned int frontbuffer_bits);
> >   void
> > @@ -112,7 +112,7 @@ bool intel_dp_get_colorimetry_status(struct intel_dp *intel_dp);
> >   int intel_dp_link_required(int pixel_clock, int bpp);
> >   int intel_dp_max_data_rate(int max_link_clock, int max_lanes);
> >   bool intel_digital_port_connected(struct intel_encoder *encoder);
> > -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv,
> > +void icl_tc_phy_disconnect(struct drm_i915_private *i915,
> >   			   struct intel_digital_port *dig_port);
> >   static inline unsigned int intel_dp_unused_lane_mask(int lane_count)
> > diff --git a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> > index 7ded95a334db..7a3bd29dcf0b 100644
> > --- a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> > +++ b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c
> > @@ -109,7 +109,7 @@ intel_dp_aux_set_backlight(const struct drm_connector_state *conn_state, u32 lev
> >    */
> >   static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&connector->encoder->base);
> >   	int freq, fxp, fxp_min, fxp_max, fxp_actual, f = 1;
> >   	u8 pn, pn_min, pn_max;
> > @@ -118,7 +118,7 @@ static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector)
> >   	 * Note that, if F x P is out of supported range, the maximum value or
> >   	 * minimum value will applied automatically. So no need to check that.
> >   	 */
> > -	freq = dev_priv->vbt.backlight.pwm_freq_hz;
> > +	freq = i915->vbt.backlight.pwm_freq_hz;
> >   	DRM_DEBUG_KMS("VBT defined backlight frequency %u Hz\n", freq);
> >   	if (!freq) {
> >   		DRM_DEBUG_KMS("Use panel default backlight frequency\n");
> > diff --git a/drivers/gpu/drm/i915/intel_dp_mst.c b/drivers/gpu/drm/i915/intel_dp_mst.c
> > index 0caf645fbbb8..ab21dc690c9f 100644
> > --- a/drivers/gpu/drm/i915/intel_dp_mst.c
> > +++ b/drivers/gpu/drm/i915/intel_dp_mst.c
> > @@ -91,7 +91,7 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder,
> >   				       struct intel_crtc_state *pipe_config,
> >   				       struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> >   	struct intel_dp *intel_dp = &intel_mst->primary->dp;
> >   	struct intel_connector *connector =
> > @@ -140,11 +140,11 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder,
> >   	pipe_config->limited_color_range =
> >   		intel_dp_limited_color_range(pipe_config, conn_state);
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		pipe_config->lane_lat_optim_mask =
> >   			bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count);
> > -	intel_ddi_compute_min_voltage_level(dev_priv, pipe_config);
> > +	intel_ddi_compute_min_voltage_level(i915, pipe_config);
> >   	return 0;
> >   }
> > @@ -285,7 +285,7 @@ static void intel_mst_pre_enable_dp(struct intel_encoder *encoder,
> >   	struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> >   	struct intel_digital_port *intel_dig_port = intel_mst->primary;
> >   	struct intel_dp *intel_dp = &intel_dig_port->dp;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = intel_dig_port->base.port;
> >   	struct intel_connector *connector =
> >   		to_intel_connector(conn_state->connector);
> > @@ -332,12 +332,12 @@ static void intel_mst_enable_dp(struct intel_encoder *encoder,
> >   	struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base);
> >   	struct intel_digital_port *intel_dig_port = intel_mst->primary;
> >   	struct intel_dp *intel_dp = &intel_dig_port->dp;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = intel_dig_port->base.port;
> >   	DRM_DEBUG_KMS("active links %d\n", intel_dp->active_mst_links);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DP_TP_STATUS(port),
> >   				    DP_TP_STATUS_ACT_SENT,
> >   				    DP_TP_STATUS_ACT_SENT,
> > @@ -495,7 +495,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
> >   	struct intel_dp *intel_dp = container_of(mgr, struct intel_dp, mst_mgr);
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> >   	struct drm_device *dev = intel_dig_port->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_connector *intel_connector;
> >   	struct drm_connector *connector;
> >   	enum pipe pipe;
> > @@ -520,7 +520,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
> >   	drm_connector_helper_add(connector, &intel_dp_mst_connector_helper_funcs);
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		struct drm_encoder *enc =
> >   			&intel_dp->mst_encoders[pipe]->base.base;
> > @@ -549,10 +549,10 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo
> >   static void intel_dp_register_mst_connector(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> > -	if (dev_priv->fbdev)
> > -		drm_fb_helper_add_one_connector(&dev_priv->fbdev->helper,
> > +	if (i915->fbdev)
> > +		drm_fb_helper_add_one_connector(&i915->fbdev->helper,
> >   						connector);
> >   	drm_connector_register(connector);
> > @@ -561,13 +561,13 @@ static void intel_dp_register_mst_connector(struct drm_connector *connector)
> >   static void intel_dp_destroy_mst_connector(struct drm_dp_mst_topology_mgr *mgr,
> >   					   struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", connector->base.id, connector->name);
> >   	drm_connector_unregister(connector);
> > -	if (dev_priv->fbdev)
> > -		drm_fb_helper_remove_one_connector(&dev_priv->fbdev->helper,
> > +	if (i915->fbdev)
> > +		drm_fb_helper_remove_one_connector(&i915->fbdev->helper,
> >   						   connector);
> >   	drm_connector_put(connector);
> > @@ -622,10 +622,10 @@ static bool
> >   intel_dp_create_fake_mst_encoders(struct intel_digital_port *intel_dig_port)
> >   {
> >   	struct intel_dp *intel_dp = &intel_dig_port->dp;
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev);
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe)
> > +	for_each_pipe(i915, pipe)
> >   		intel_dp->mst_encoders[pipe] = intel_dp_create_fake_mst_encoder(intel_dig_port, pipe);
> >   	return true;
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.c b/drivers/gpu/drm/i915/intel_dpio_phy.c
> > index bdbe41759827..c85672bb1a7d 100644
> > --- a/drivers/gpu/drm/i915/intel_dpio_phy.c
> > +++ b/drivers/gpu/drm/i915/intel_dpio_phy.c
> > @@ -212,9 +212,9 @@ static const struct bxt_ddi_phy_info glk_ddi_phy_info[] = {
> >   };
> >   static const struct bxt_ddi_phy_info *
> > -bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count)
> > +bxt_get_phy_list(struct drm_i915_private *i915, int *count)
> >   {
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		*count =  ARRAY_SIZE(glk_ddi_phy_info);
> >   		return glk_ddi_phy_info;
> >   	} else {
> > @@ -224,22 +224,22 @@ bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count)
> >   }
> >   static const struct bxt_ddi_phy_info *
> > -bxt_get_phy_info(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> > +bxt_get_phy_info(struct drm_i915_private *i915, enum dpio_phy phy)
> >   {
> >   	int count;
> >   	const struct bxt_ddi_phy_info *phy_list =
> > -		bxt_get_phy_list(dev_priv, &count);
> > +		bxt_get_phy_list(i915, &count);
> >   	return &phy_list[phy];
> >   }
> > -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> > +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port,
> >   			     enum dpio_phy *phy, enum dpio_channel *ch)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info, *phys;
> >   	int i, count;
> > -	phys = bxt_get_phy_list(dev_priv, &count);
> > +	phys = bxt_get_phy_list(i915, &count);
> >   	for (i = 0; i < count; i++) {
> >   		phy_info = &phys[i];
> > @@ -263,7 +263,7 @@ void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> >   	*ch = DPIO_CH0;
> >   }
> > -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> > +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915,
> >   				  enum port port, u32 margin, u32 scale,
> >   				  u32 enable, u32 deemphasis)
> >   {
> > @@ -271,7 +271,7 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> >   	enum dpio_phy phy;
> >   	enum dpio_channel ch;
> > -	bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> > +	bxt_port_to_phy_channel(i915, port, &phy, &ch);
> >   	/*
> >   	 * While we write to the group register to program all lanes at once we
> > @@ -306,12 +306,12 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(BXT_PORT_PCS_DW10_GRP(phy, ch), val);
> >   }
> > -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> > +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915,
> >   			    enum dpio_phy phy)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info;
> > -	phy_info = bxt_get_phy_info(dev_priv, phy);
> > +	phy_info = bxt_get_phy_info(i915, phy);
> >   	if (!(I915_READ(BXT_P_CR_GT_DISP_PWRON) & phy_info->pwron_mask))
> >   		return false;
> > @@ -334,37 +334,37 @@ bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static u32 bxt_get_grc(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> > +static u32 bxt_get_grc(struct drm_i915_private *i915, enum dpio_phy phy)
> >   {
> >   	u32 val = I915_READ(BXT_PORT_REF_DW6(phy));
> >   	return (val & GRC_CODE_MASK) >> GRC_CODE_SHIFT;
> >   }
> > -static void bxt_phy_wait_grc_done(struct drm_i915_private *dev_priv,
> > +static void bxt_phy_wait_grc_done(struct drm_i915_private *i915,
> >   				  enum dpio_phy phy)
> >   {
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    BXT_PORT_REF_DW3(phy),
> >   				    GRC_DONE, GRC_DONE,
> >   				    10))
> >   		DRM_ERROR("timeout waiting for PHY%d GRC\n", phy);
> >   }
> > -static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> > +static void _bxt_ddi_phy_init(struct drm_i915_private *i915,
> >   			      enum dpio_phy phy)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info;
> >   	u32 val;
> > -	phy_info = bxt_get_phy_info(dev_priv, phy);
> > +	phy_info = bxt_get_phy_info(i915, phy);
> > -	if (bxt_ddi_phy_is_enabled(dev_priv, phy)) {
> > +	if (bxt_ddi_phy_is_enabled(i915, phy)) {
> >   		/* Still read out the GRC value for state verification */
> >   		if (phy_info->rcomp_phy != -1)
> > -			dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv, phy);
> > +			i915->bxt_phy_grc = bxt_get_grc(i915, phy);
> > -		if (bxt_ddi_phy_verify_state(dev_priv, phy)) {
> > +		if (bxt_ddi_phy_verify_state(i915, phy)) {
> >   			DRM_DEBUG_DRIVER("DDI PHY %d already enabled, "
> >   					 "won't reprogram it\n", phy);
> >   			return;
> > @@ -386,7 +386,7 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> >   	 * The flag should get set in 100us according to the HW team, but
> >   	 * use 1ms due to occasional timeouts observed with that.
> >   	 */
> > -	if (intel_wait_for_register_fw(&dev_priv->uncore,
> > +	if (intel_wait_for_register_fw(&i915->uncore,
> >   				       BXT_PORT_CL1CM_DW0(phy),
> >   				       PHY_RESERVED | PHY_POWER_GOOD,
> >   				       PHY_POWER_GOOD,
> > @@ -419,14 +419,14 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> >   	if (phy_info->rcomp_phy != -1) {
> >   		u32 grc_code;
> > -		bxt_phy_wait_grc_done(dev_priv, phy_info->rcomp_phy);
> > +		bxt_phy_wait_grc_done(i915, phy_info->rcomp_phy);
> >   		/*
> >   		 * PHY0 isn't connected to an RCOMP resistor so copy over
> >   		 * the corresponding calibrated value from PHY1, and disable
> >   		 * the automatic calibration on PHY0.
> >   		 */
> > -		val = dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv,
> > +		val = i915->bxt_phy_grc = bxt_get_grc(i915,
> >   							  phy_info->rcomp_phy);
> >   		grc_code = val << GRC_CODE_FAST_SHIFT |
> >   			   val << GRC_CODE_SLOW_SHIFT |
> > @@ -446,12 +446,12 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(BXT_PHY_CTL_FAMILY(phy), val);
> >   }
> > -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> > +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info;
> >   	u32 val;
> > -	phy_info = bxt_get_phy_info(dev_priv, phy);
> > +	phy_info = bxt_get_phy_info(i915, phy);
> >   	val = I915_READ(BXT_PHY_CTL_FAMILY(phy));
> >   	val &= ~COMMON_RESET_DIS;
> > @@ -462,34 +462,34 @@ void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> >   	I915_WRITE(BXT_P_CR_GT_DISP_PWRON, val);
> >   }
> > -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy)
> > +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info =
> > -		bxt_get_phy_info(dev_priv, phy);
> > +		bxt_get_phy_info(i915, phy);
> >   	enum dpio_phy rcomp_phy = phy_info->rcomp_phy;
> >   	bool was_enabled;
> > -	lockdep_assert_held(&dev_priv->power_domains.lock);
> > +	lockdep_assert_held(&i915->power_domains.lock);
> >   	was_enabled = true;
> >   	if (rcomp_phy != -1)
> > -		was_enabled = bxt_ddi_phy_is_enabled(dev_priv, rcomp_phy);
> > +		was_enabled = bxt_ddi_phy_is_enabled(i915, rcomp_phy);
> >   	/*
> >   	 * We need to copy the GRC calibration value from rcomp_phy,
> >   	 * so make sure it's powered up.
> >   	 */
> >   	if (!was_enabled)
> > -		_bxt_ddi_phy_init(dev_priv, rcomp_phy);
> > +		_bxt_ddi_phy_init(i915, rcomp_phy);
> > -	_bxt_ddi_phy_init(dev_priv, phy);
> > +	_bxt_ddi_phy_init(i915, phy);
> >   	if (!was_enabled)
> > -		bxt_ddi_phy_uninit(dev_priv, rcomp_phy);
> > +		bxt_ddi_phy_uninit(i915, rcomp_phy);
> >   }
> >   static bool __printf(6, 7)
> > -__phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> > +__phy_reg_verify_state(struct drm_i915_private *i915, enum dpio_phy phy,
> >   		       i915_reg_t reg, u32 mask, u32 expected,
> >   		       const char *reg_fmt, ...)
> >   {
> > @@ -515,20 +515,20 @@ __phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy,
> >   	return false;
> >   }
> > -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> > +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915,
> >   			      enum dpio_phy phy)
> >   {
> >   	const struct bxt_ddi_phy_info *phy_info;
> >   	u32 mask;
> >   	bool ok;
> > -	phy_info = bxt_get_phy_info(dev_priv, phy);
> > +	phy_info = bxt_get_phy_info(i915, phy);
> >   #define _CHK(reg, mask, exp, fmt, ...)					\
> > -	__phy_reg_verify_state(dev_priv, phy, reg, mask, exp, fmt,	\
> > +	__phy_reg_verify_state(i915, phy, reg, mask, exp, fmt,	\
> >   			       ## __VA_ARGS__)
> > -	if (!bxt_ddi_phy_is_enabled(dev_priv, phy))
> > +	if (!bxt_ddi_phy_is_enabled(i915, phy))
> >   		return false;
> >   	ok = true;
> > @@ -552,7 +552,7 @@ bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> >   			   "BXT_PORT_CL2CM_DW6(%d)", phy);
> >   	if (phy_info->rcomp_phy != -1) {
> > -		u32 grc_code = dev_priv->bxt_phy_grc;
> > +		u32 grc_code = i915->bxt_phy_grc;
> >   		grc_code = grc_code << GRC_CODE_FAST_SHIFT |
> >   			   grc_code << GRC_CODE_SLOW_SHIFT |
> > @@ -591,13 +591,13 @@ bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count)
> >   void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> >   				     u8 lane_lat_optim_mask)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	enum dpio_phy phy;
> >   	enum dpio_channel ch;
> >   	int lane;
> > -	bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> > +	bxt_port_to_phy_channel(i915, port, &phy, &ch);
> >   	for (lane = 0; lane < 4; lane++) {
> >   		u32 val = I915_READ(BXT_PORT_TX_DW14_LN(phy, ch, lane));
> > @@ -617,14 +617,14 @@ void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> >   u8
> >   bxt_ddi_phy_get_lane_lat_optim_mask(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum port port = encoder->port;
> >   	enum dpio_phy phy;
> >   	enum dpio_channel ch;
> >   	int lane;
> >   	u8 mask;
> > -	bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> > +	bxt_port_to_phy_channel(i915, port, &phy, &ch);
> >   	mask = 0;
> >   	for (lane = 0; lane < 4; lane++) {
> > @@ -642,7 +642,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> >   			      u32 deemph_reg_value, u32 margin_reg_value,
> >   			      bool uniq_trans_scale)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc);
> >   	enum dpio_channel ch = vlv_dport_to_channel(dport);
> > @@ -650,46 +650,46 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> >   	u32 val;
> >   	int i;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Clear calc init */
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch));
> >   	val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3);
> >   	val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK);
> >   	val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val);
> >   	if (intel_crtc->config->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch));
> >   		val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3);
> >   		val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK);
> >   		val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val);
> >   	}
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW9(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW9(ch));
> >   	val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK);
> >   	val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW9(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW9(ch), val);
> >   	if (intel_crtc->config->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW9(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW9(ch));
> >   		val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK);
> >   		val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW9(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW9(ch), val);
> >   	}
> >   	/* Program swing deemph */
> >   	for (i = 0; i < intel_crtc->config->lane_count; i++) {
> > -		val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW4(ch, i));
> > +		val = vlv_dpio_read(i915, pipe, CHV_TX_DW4(ch, i));
> >   		val &= ~DPIO_SWING_DEEMPH9P5_MASK;
> >   		val |= deemph_reg_value << DPIO_SWING_DEEMPH9P5_SHIFT;
> > -		vlv_dpio_write(dev_priv, pipe, CHV_TX_DW4(ch, i), val);
> > +		vlv_dpio_write(i915, pipe, CHV_TX_DW4(ch, i), val);
> >   	}
> >   	/* Program swing margin */
> >   	for (i = 0; i < intel_crtc->config->lane_count; i++) {
> > -		val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW2(ch, i));
> > +		val = vlv_dpio_read(i915, pipe, CHV_TX_DW2(ch, i));
> >   		val &= ~DPIO_SWING_MARGIN000_MASK;
> >   		val |= margin_reg_value << DPIO_SWING_MARGIN000_SHIFT;
> > @@ -702,7 +702,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> >   		val &= ~(0xff << DPIO_UNIQ_TRANS_SCALE_SHIFT);
> >   		val |= 0x9a << DPIO_UNIQ_TRANS_SCALE_SHIFT;
> > -		vlv_dpio_write(dev_priv, pipe, CHV_TX_DW2(ch, i), val);
> > +		vlv_dpio_write(i915, pipe, CHV_TX_DW2(ch, i), val);
> >   	}
> >   	/*
> > @@ -712,70 +712,70 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder,
> >   	 * 27 for ch0 and ch1.
> >   	 */
> >   	for (i = 0; i < intel_crtc->config->lane_count; i++) {
> > -		val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW3(ch, i));
> > +		val = vlv_dpio_read(i915, pipe, CHV_TX_DW3(ch, i));
> >   		if (uniq_trans_scale)
> >   			val |= DPIO_TX_UNIQ_TRANS_SCALE_EN;
> >   		else
> >   			val &= ~DPIO_TX_UNIQ_TRANS_SCALE_EN;
> > -		vlv_dpio_write(dev_priv, pipe, CHV_TX_DW3(ch, i), val);
> > +		vlv_dpio_write(i915, pipe, CHV_TX_DW3(ch, i), val);
> >   	}
> >   	/* Start swing calculation */
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch));
> >   	val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val);
> >   	if (intel_crtc->config->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch));
> >   		val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val);
> >   	}
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void chv_data_lane_soft_reset(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *crtc_state,
> >   			      bool reset)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base));
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW0(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW0(ch));
> >   	if (reset)
> >   		val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET);
> >   	else
> >   		val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW0(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW0(ch), val);
> >   	if (crtc_state->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW0(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW0(ch));
> >   		if (reset)
> >   			val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET);
> >   		else
> >   			val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW0(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW0(ch), val);
> >   	}
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW1(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW1(ch));
> >   	val |= CHV_PCS_REQ_SOFTRESET_EN;
> >   	if (reset)
> >   		val &= ~DPIO_PCS_CLK_SOFT_RESET;
> >   	else
> >   		val |= DPIO_PCS_CLK_SOFT_RESET;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW1(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW1(ch), val);
> >   	if (crtc_state->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW1(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW1(ch));
> >   		val |= CHV_PCS_REQ_SOFTRESET_EN;
> >   		if (reset)
> >   			val &= ~DPIO_PCS_CLK_SOFT_RESET;
> >   		else
> >   			val |= DPIO_PCS_CLK_SOFT_RESET;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW1(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW1(ch), val);
> >   	}
> >   }
> > @@ -783,7 +783,7 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> >   			    const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum dpio_channel ch = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -797,51 +797,51 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> >   	 */
> >   	if (ch == DPIO_CH0 && pipe == PIPE_B)
> >   		dport->release_cl2_override =
> > -			!chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, true);
> > +			!chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, true);
> >   	chv_phy_powergate_lanes(encoder, true, lane_mask);
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Assert data lane reset */
> >   	chv_data_lane_soft_reset(encoder, crtc_state, true);
> >   	/* program left/right clock distribution */
> >   	if (pipe != PIPE_B) {
> > -		val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0);
> > +		val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0);
> >   		val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK);
> >   		if (ch == DPIO_CH0)
> >   			val |= CHV_BUFLEFTENA1_FORCE;
> >   		if (ch == DPIO_CH1)
> >   			val |= CHV_BUFRIGHTENA1_FORCE;
> > -		vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val);
> > +		vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val);
> >   	} else {
> > -		val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1);
> > +		val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1);
> >   		val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK);
> >   		if (ch == DPIO_CH0)
> >   			val |= CHV_BUFLEFTENA2_FORCE;
> >   		if (ch == DPIO_CH1)
> >   			val |= CHV_BUFRIGHTENA2_FORCE;
> > -		vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val);
> > +		vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val);
> >   	}
> >   	/* program clock channel usage */
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(ch));
> >   	val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE;
> >   	if (pipe != PIPE_B)
> >   		val &= ~CHV_PCS_USEDCLKCHANNEL;
> >   	else
> >   		val |= CHV_PCS_USEDCLKCHANNEL;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW8(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW8(ch), val);
> >   	if (crtc_state->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW8(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW8(ch));
> >   		val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE;
> >   		if (pipe != PIPE_B)
> >   			val &= ~CHV_PCS_USEDCLKCHANNEL;
> >   		else
> >   			val |= CHV_PCS_USEDCLKCHANNEL;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW8(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW8(ch), val);
> >   	}
> >   	/*
> > @@ -849,14 +849,14 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder,
> >   	 * matches the pipe, but here we need to
> >   	 * pick the CL based on the port.
> >   	 */
> > -	val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW19(ch));
> > +	val = vlv_dpio_read(i915, pipe, CHV_CMN_DW19(ch));
> >   	if (pipe != PIPE_B)
> >   		val &= ~CHV_CMN_USEDCLKCHANNEL;
> >   	else
> >   		val |= CHV_CMN_USEDCLKCHANNEL;
> > -	vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW19(ch), val);
> > +	vlv_dpio_write(i915, pipe, CHV_CMN_DW19(ch), val);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> > @@ -864,24 +864,24 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   {
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum dpio_channel ch = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = crtc->pipe;
> >   	int data, i, stagger;
> >   	u32 val;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* allow hardware to manage TX FIFO reset source */
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch));
> >   	val &= ~DPIO_LANEDESKEW_STRAP_OVRD;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val);
> >   	if (crtc_state->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch));
> >   		val &= ~DPIO_LANEDESKEW_STRAP_OVRD;
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val);
> >   	}
> >   	/* Program Tx lane latency optimal setting*/
> > @@ -891,7 +891,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   			data = 0x0;
> >   		else
> >   			data = (i == 1) ? 0x0 : 0x1;
> > -		vlv_dpio_write(dev_priv, pipe, CHV_TX_DW14(ch, i),
> > +		vlv_dpio_write(i915, pipe, CHV_TX_DW14(ch, i),
> >   				data << DPIO_UPAR_SHIFT);
> >   	}
> > @@ -907,17 +907,17 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   	else
> >   		stagger = 0x2;
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch));
> >   	val |= DPIO_TX2_STAGGER_MASK(0x1f);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val);
> >   	if (crtc_state->lane_count > 2) {
> > -		val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch));
> > +		val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch));
> >   		val |= DPIO_TX2_STAGGER_MASK(0x1f);
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val);
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val);
> >   	}
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW12(ch),
> > +	vlv_dpio_write(i915, pipe, VLV_PCS01_DW12(ch),
> >   		       DPIO_LANESTAGGER_STRAP(stagger) |
> >   		       DPIO_LANESTAGGER_STRAP_OVRD |
> >   		       DPIO_TX1_STAGGER_MASK(0x1f) |
> > @@ -925,7 +925,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   		       DPIO_TX2_STAGGER_MULT(0));
> >   	if (crtc_state->lane_count > 2) {
> > -		vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW12(ch),
> > +		vlv_dpio_write(i915, pipe, VLV_PCS23_DW12(ch),
> >   			       DPIO_LANESTAGGER_STRAP(stagger) |
> >   			       DPIO_LANESTAGGER_STRAP_OVRD |
> >   			       DPIO_TX1_STAGGER_MASK(0x1f) |
> > @@ -936,16 +936,16 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   	/* Deassert data lane reset */
> >   	chv_data_lane_soft_reset(encoder, crtc_state, false);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void chv_phy_release_cl2_override(struct intel_encoder *encoder)
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	if (dport->release_cl2_override) {
> > -		chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, false);
> > +		chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, false);
> >   		dport->release_cl2_override = false;
> >   	}
> >   }
> > @@ -953,24 +953,24 @@ void chv_phy_release_cl2_override(struct intel_encoder *encoder)
> >   void chv_phy_post_pll_disable(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *old_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum pipe pipe = to_intel_crtc(old_crtc_state->base.crtc)->pipe;
> >   	u32 val;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* disable left/right clock distribution */
> >   	if (pipe != PIPE_B) {
> > -		val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0);
> > +		val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0);
> >   		val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK);
> > -		vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val);
> > +		vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val);
> >   	} else {
> > -		val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1);
> > +		val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1);
> >   		val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK);
> > -		vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val);
> > +		vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val);
> >   	}
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   	/*
> >   	 * Leave the power down bit cleared for at least one
> > @@ -988,57 +988,57 @@ void vlv_set_phy_signal_level(struct intel_encoder *encoder,
> >   			      u32 demph_reg_value, u32 preemph_reg_value,
> >   			      u32 uniqtranscale_reg_value, u32 tx3_demph)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc);
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> >   	enum dpio_channel port = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = intel_crtc->pipe;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), 0x00000000);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW4(port), demph_reg_value);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW2(port),
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), 0x00000000);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW4(port), demph_reg_value);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW2(port),
> >   			 uniqtranscale_reg_value);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW3(port), 0x0C782040);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW3(port), 0x0C782040);
> >   	if (tx3_demph)
> > -		vlv_dpio_write(dev_priv, pipe, VLV_TX3_DW4(port), tx3_demph);
> > +		vlv_dpio_write(i915, pipe, VLV_TX3_DW4(port), tx3_demph);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW11(port), 0x00030000);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW9(port), preemph_reg_value);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW11(port), 0x00030000);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW9(port), preemph_reg_value);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void vlv_phy_pre_pll_enable(struct intel_encoder *encoder,
> >   			    const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum dpio_channel port = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = crtc->pipe;
> >   	/* Program Tx lane resets to default */
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port),
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port),
> >   			 DPIO_PCS_TX_LANE2_RESET |
> >   			 DPIO_PCS_TX_LANE1_RESET);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port),
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port),
> >   			 DPIO_PCS_CLK_CRI_RXEB_EIOS_EN |
> >   			 DPIO_PCS_CLK_CRI_RXDIGFILTSG_EN |
> >   			 (1<<DPIO_PCS_CLK_DATAWIDTH_SHIFT) |
> >   				 DPIO_PCS_CLK_SOFT_RESET);
> >   	/* Fix up inter-pair skew failure */
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW12(port), 0x00750f00);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW11(port), 0x00001500);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_TX_DW14(port), 0x40400000);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW12(port), 0x00750f00);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW11(port), 0x00001500);
> > +	vlv_dpio_write(i915, pipe, VLV_TX_DW14(port), 0x40400000);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> > @@ -1046,42 +1046,42 @@ void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder,
> >   {
> >   	struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base);
> >   	struct intel_digital_port *dport = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	enum dpio_channel port = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 val;
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Enable clock channels for this port */
> > -	val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(port));
> > +	val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(port));
> >   	val = 0;
> >   	if (pipe)
> >   		val |= (1<<21);
> >   	else
> >   		val &= ~(1<<21);
> >   	val |= 0x001000c4;
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW8(port), val);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW8(port), val);
> >   	/* Program lane clock */
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW14(port), 0x00760018);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW23(port), 0x00400888);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW14(port), 0x00760018);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW23(port), 0x00400888);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   void vlv_phy_reset_lanes(struct intel_encoder *encoder,
> >   			 const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	enum dpio_channel port = vlv_dport_to_channel(dport);
> >   	enum pipe pipe = crtc->pipe;
> > -	vlv_dpio_get(dev_priv);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port), 0x00000000);
> > -	vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port), 0x00e00060);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_get(i915);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port), 0x00000000);
> > +	vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port), 0x00e00060);
> > +	vlv_dpio_put(i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.h b/drivers/gpu/drm/i915/intel_dpio_phy.h
> > index f418aab90b7e..d4c9393bebce 100644
> > --- a/drivers/gpu/drm/i915/intel_dpio_phy.h
> > +++ b/drivers/gpu/drm/i915/intel_dpio_phy.h
> > @@ -15,16 +15,16 @@ struct drm_i915_private;
> >   struct intel_crtc_state;
> >   struct intel_encoder;
> > -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port,
> > +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port,
> >   			     enum dpio_phy *phy, enum dpio_channel *ch);
> > -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv,
> > +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915,
> >   				  enum port port, u32 margin, u32 scale,
> >   				  u32 enable, u32 deemphasis);
> > -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy);
> > -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy);
> > -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv,
> > +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy);
> > +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy);
> > +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915,
> >   			    enum dpio_phy phy);
> > -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv,
> > +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915,
> >   			      enum dpio_phy phy);
> >   u8 bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count);
> >   void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder,
> > diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.c b/drivers/gpu/drm/i915/intel_dpll_mgr.c
> > index 2d4e7b9a7b9d..2dabcdf27df4 100644
> > --- a/drivers/gpu/drm/i915/intel_dpll_mgr.c
> > +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.c
> > @@ -45,14 +45,14 @@
> >    */
> >   static void
> > -intel_atomic_duplicate_dpll_state(struct drm_i915_private *dev_priv,
> > +intel_atomic_duplicate_dpll_state(struct drm_i915_private *i915,
> >   				  struct intel_shared_dpll_state *shared_dpll)
> >   {
> >   	enum intel_dpll_id i;
> >   	/* Copy shared dpll state */
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > -		struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i];
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> > +		struct intel_shared_dpll *pll = &i915->shared_dplls[i];
> >   		shared_dpll[i] = pll->state;
> >   	}
> > @@ -77,40 +77,40 @@ intel_atomic_get_shared_dpll_state(struct drm_atomic_state *s)
> >   /**
> >    * intel_get_shared_dpll_by_id - get a DPLL given its id
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @id: pll id
> >    *
> >    * Returns:
> >    * A pointer to the DPLL with @id
> >    */
> >   struct intel_shared_dpll *
> > -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv,
> > +intel_get_shared_dpll_by_id(struct drm_i915_private *i915,
> >   			    enum intel_dpll_id id)
> >   {
> > -	return &dev_priv->shared_dplls[id];
> > +	return &i915->shared_dplls[id];
> >   }
> >   /**
> >    * intel_get_shared_dpll_id - get the id of a DPLL
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @pll: the DPLL
> >    *
> >    * Returns:
> >    * The id of @pll
> >    */
> >   enum intel_dpll_id
> > -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv,
> > +intel_get_shared_dpll_id(struct drm_i915_private *i915,
> >   			 struct intel_shared_dpll *pll)
> >   {
> > -	if (WARN_ON(pll < dev_priv->shared_dplls||
> > -		    pll > &dev_priv->shared_dplls[dev_priv->num_shared_dpll]))
> > +	if (WARN_ON(pll < i915->shared_dplls||
> > +		    pll > &i915->shared_dplls[i915->num_shared_dpll]))
> >   		return -1;
> > -	return (enum intel_dpll_id) (pll - dev_priv->shared_dplls);
> > +	return (enum intel_dpll_id) (pll - i915->shared_dplls);
> >   }
> >   /* For ILK+ */
> > -void assert_shared_dpll(struct drm_i915_private *dev_priv,
> > +void assert_shared_dpll(struct drm_i915_private *i915,
> >   			struct intel_shared_dpll *pll,
> >   			bool state)
> >   {
> > @@ -120,7 +120,7 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv,
> >   	if (WARN(!pll, "asserting DPLL %s with no DPLL\n", onoff(state)))
> >   		return;
> > -	cur_state = pll->info->funcs->get_hw_state(dev_priv, pll, &hw_state);
> > +	cur_state = pll->info->funcs->get_hw_state(i915, pll, &hw_state);
> >   	I915_STATE_WARN(cur_state != state,
> >   	     "%s assertion failure (expected %s, current %s)\n",
> >   			pll->info->name, onoff(state), onoff(cur_state));
> > @@ -136,22 +136,22 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv,
> >   void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> >   	if (WARN_ON(pll == NULL))
> >   		return;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	WARN_ON(!pll->state.crtc_mask);
> >   	if (!pll->active_mask) {
> >   		DRM_DEBUG_DRIVER("setting up %s\n", pll->info->name);
> >   		WARN_ON(pll->on);
> > -		assert_shared_dpll_disabled(dev_priv, pll);
> > +		assert_shared_dpll_disabled(i915, pll);
> > -		pll->info->funcs->prepare(dev_priv, pll);
> > +		pll->info->funcs->prepare(i915, pll);
> >   	}
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   /**
> > @@ -163,7 +163,7 @@ void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> >   	unsigned int crtc_mask = drm_crtc_mask(&crtc->base);
> >   	unsigned int old_mask;
> > @@ -171,7 +171,7 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   	if (WARN_ON(pll == NULL))
> >   		return;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	old_mask = pll->active_mask;
> >   	if (WARN_ON(!(pll->state.crtc_mask & crtc_mask)) ||
> > @@ -186,17 +186,17 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   	if (old_mask) {
> >   		WARN_ON(!pll->on);
> > -		assert_shared_dpll_enabled(dev_priv, pll);
> > +		assert_shared_dpll_enabled(i915, pll);
> >   		goto out;
> >   	}
> >   	WARN_ON(pll->on);
> >   	DRM_DEBUG_KMS("enabling %s\n", pll->info->name);
> > -	pll->info->funcs->enable(dev_priv, pll);
> > +	pll->info->funcs->enable(i915, pll);
> >   	pll->on = true;
> >   out:
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   /**
> > @@ -208,18 +208,18 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll = crtc_state->shared_dpll;
> >   	unsigned int crtc_mask = drm_crtc_mask(&crtc->base);
> >   	/* PCH only available on ILK+ */
> > -	if (INTEL_GEN(dev_priv) < 5)
> > +	if (INTEL_GEN(i915) < 5)
> >   		return;
> >   	if (pll == NULL)
> >   		return;
> > -	mutex_lock(&dev_priv->dpll_lock);
> > +	mutex_lock(&i915->dpll_lock);
> >   	if (WARN_ON(!(pll->active_mask & crtc_mask)))
> >   		goto out;
> > @@ -227,7 +227,7 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   		      pll->info->name, pll->active_mask, pll->on,
> >   		      crtc->base.base.id);
> > -	assert_shared_dpll_enabled(dev_priv, pll);
> > +	assert_shared_dpll_enabled(i915, pll);
> >   	WARN_ON(!pll->on);
> >   	pll->active_mask &= ~crtc_mask;
> > @@ -235,11 +235,11 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state)
> >   		goto out;
> >   	DRM_DEBUG_KMS("disabling %s\n", pll->info->name);
> > -	pll->info->funcs->disable(dev_priv, pll);
> > +	pll->info->funcs->disable(i915, pll);
> >   	pll->on = false;
> >   out:
> > -	mutex_unlock(&dev_priv->dpll_lock);
> > +	mutex_unlock(&i915->dpll_lock);
> >   }
> >   static struct intel_shared_dpll *
> > @@ -248,7 +248,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state,
> >   		       enum intel_dpll_id range_max)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll, *unused_pll = NULL;
> >   	struct intel_shared_dpll_state *shared_dpll;
> >   	enum intel_dpll_id i;
> > @@ -256,7 +256,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state,
> >   	shared_dpll = intel_atomic_get_shared_dpll_state(crtc_state->base.state);
> >   	for (i = range_min; i <= range_max; i++) {
> > -		pll = &dev_priv->shared_dplls[i];
> > +		pll = &i915->shared_dplls[i];
> >   		/* Only want to check enabled timings first */
> >   		if (shared_dpll[i].crtc_mask == 0) {
> > @@ -322,7 +322,7 @@ intel_reference_shared_dpll(struct intel_shared_dpll *pll,
> >    */
> >   void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->dev);
> > +	struct drm_i915_private *i915 = to_i915(state->dev);
> >   	struct intel_shared_dpll_state *shared_dpll;
> >   	struct intel_shared_dpll *pll;
> >   	enum intel_dpll_id i;
> > @@ -331,10 +331,10 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> >   		return;
> >   	shared_dpll = to_intel_atomic_state(state)->shared_dpll;
> > -	for (i = 0; i < dev_priv->num_shared_dpll; i++) {
> > +	for (i = 0; i < i915->num_shared_dpll; i++) {
> >   		struct intel_shared_dpll_state tmp;
> > -		pll = &dev_priv->shared_dplls[i];
> > +		pll = &i915->shared_dplls[i];
> >   		tmp = pll->state;
> >   		pll->state = shared_dpll[i];
> > @@ -342,7 +342,7 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state)
> >   	}
> >   }
> > -static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *i915,
> >   				      struct intel_shared_dpll *pll,
> >   				      struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -350,7 +350,7 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	intel_wakeref_t wakeref;
> >   	u32 val;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -360,12 +360,12 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	hw_state->fp0 = I915_READ(PCH_FP0(id));
> >   	hw_state->fp1 = I915_READ(PCH_FP1(id));
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return val & DPLL_VCO_ENABLE;
> >   }
> > -static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv,
> > +static void ibx_pch_dpll_prepare(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -374,12 +374,12 @@ static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(PCH_FP1(id), pll->state.hw_state.fp1);
> >   }
> > -static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv)
> > +static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> >   	bool enabled;
> > -	I915_STATE_WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)));
> > +	I915_STATE_WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)));
> >   	val = I915_READ(PCH_DREF_CONTROL);
> >   	enabled = !!(val & (DREF_SSC_SOURCE_MASK | DREF_NONSPREAD_SOURCE_MASK |
> > @@ -387,13 +387,13 @@ static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv)
> >   	I915_STATE_WARN(!enabled, "PCH refclk assertion failure, should be active but is disabled\n");
> >   }
> > -static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv,
> > +static void ibx_pch_dpll_enable(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> >   	/* PCH refclock must be enabled first */
> > -	ibx_assert_pch_refclk_enabled(dev_priv);
> > +	ibx_assert_pch_refclk_enabled(i915);
> >   	I915_WRITE(PCH_DPLL(id), pll->state.hw_state.dpll);
> > @@ -411,7 +411,7 @@ static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv,
> >   	udelay(200);
> >   }
> > -static void ibx_pch_dpll_disable(struct drm_i915_private *dev_priv,
> > +static void ibx_pch_dpll_disable(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -426,14 +426,14 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state,
> >   	     struct intel_encoder *encoder)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll;
> >   	enum intel_dpll_id i;
> > -	if (HAS_PCH_IBX(dev_priv)) {
> > +	if (HAS_PCH_IBX(i915)) {
> >   		/* Ironlake PCH has a fixed PLL->PCH pipe mapping. */
> >   		i = (enum intel_dpll_id) crtc->pipe;
> > -		pll = &dev_priv->shared_dplls[i];
> > +		pll = &i915->shared_dplls[i];
> >   		DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n",
> >   			      crtc->base.base.id, crtc->base.name,
> > @@ -453,7 +453,7 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static void ibx_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void ibx_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: dpll: 0x%x, dpll_md: 0x%x, "
> > @@ -471,7 +471,7 @@ static const struct intel_shared_dpll_funcs ibx_pch_dpll_funcs = {
> >   	.get_hw_state = ibx_pch_dpll_get_hw_state,
> >   };
> > -static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_wrpll_enable(struct drm_i915_private *i915,
> >   			       struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -481,7 +481,7 @@ static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv,
> >   	udelay(20);
> >   }
> > -static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_spll_enable(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll)
> >   {
> >   	I915_WRITE(SPLL_CTL, pll->state.hw_state.spll);
> > @@ -489,7 +489,7 @@ static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv,
> >   	udelay(20);
> >   }
> > -static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_wrpll_disable(struct drm_i915_private *i915,
> >   				  struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -500,7 +500,7 @@ static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(WRPLL_CTL(id));
> >   }
> > -static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_spll_disable(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll)
> >   {
> >   	u32 val;
> > @@ -510,7 +510,7 @@ static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(SPLL_CTL);
> >   }
> > -static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *i915,
> >   				       struct intel_shared_dpll *pll,
> >   				       struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -518,7 +518,7 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	intel_wakeref_t wakeref;
> >   	u32 val;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -526,19 +526,19 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(WRPLL_CTL(id));
> >   	hw_state->wrpll = val;
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return val & WRPLL_PLL_ENABLE;
> >   }
> > -static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *i915,
> >   				      struct intel_shared_dpll *pll,
> >   				      struct intel_dpll_hw_state *hw_state)
> >   {
> >   	intel_wakeref_t wakeref;
> >   	u32 val;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -546,7 +546,7 @@ static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(SPLL_CTL);
> >   	hw_state->spll = val;
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return val & SPLL_PLL_ENABLE;
> >   }
> > @@ -793,7 +793,7 @@ static struct intel_shared_dpll *hsw_ddi_hdmi_get_dpll(struct intel_crtc_state *
> >   static struct intel_shared_dpll *
> >   hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_shared_dpll *pll;
> >   	enum intel_dpll_id pll_id;
> >   	int clock = crtc_state->port_clock;
> > @@ -813,7 +813,7 @@ hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state)
> >   		return NULL;
> >   	}
> > -	pll = intel_get_shared_dpll_by_id(dev_priv, pll_id);
> > +	pll = intel_get_shared_dpll_by_id(i915, pll_id);
> >   	if (!pll)
> >   		return NULL;
> > @@ -855,7 +855,7 @@ hsw_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static void hsw_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void hsw_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: wrpll: 0x%x spll: 0x%x\n",
> > @@ -874,17 +874,17 @@ static const struct intel_shared_dpll_funcs hsw_ddi_spll_funcs = {
> >   	.get_hw_state = hsw_ddi_spll_get_hw_state,
> >   };
> > -static void hsw_ddi_lcpll_enable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_lcpll_enable(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll)
> >   {
> >   }
> > -static void hsw_ddi_lcpll_disable(struct drm_i915_private *dev_priv,
> > +static void hsw_ddi_lcpll_disable(struct drm_i915_private *i915,
> >   				  struct intel_shared_dpll *pll)
> >   {
> >   }
> > -static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *i915,
> >   				       struct intel_shared_dpll *pll,
> >   				       struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -928,7 +928,7 @@ static const struct skl_dpll_regs skl_dpll_regs[4] = {
> >   	},
> >   };
> > -static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv,
> > +static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *i915,
> >   				    struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -945,13 +945,13 @@ static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(DPLL_CTRL1);
> >   }
> > -static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> > +static void skl_ddi_pll_enable(struct drm_i915_private *i915,
> >   			       struct intel_shared_dpll *pll)
> >   {
> >   	const struct skl_dpll_regs *regs = skl_dpll_regs;
> >   	const enum intel_dpll_id id = pll->info->id;
> > -	skl_ddi_pll_write_ctrl1(dev_priv, pll);
> > +	skl_ddi_pll_write_ctrl1(i915, pll);
> >   	I915_WRITE(regs[id].cfgcr1, pll->state.hw_state.cfgcr1);
> >   	I915_WRITE(regs[id].cfgcr2, pll->state.hw_state.cfgcr2);
> > @@ -962,7 +962,7 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(regs[id].ctl,
> >   		   I915_READ(regs[id].ctl) | LCPLL_PLL_ENABLE);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    DPLL_STATUS,
> >   				    DPLL_LOCK(id),
> >   				    DPLL_LOCK(id),
> > @@ -970,13 +970,13 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   		DRM_ERROR("DPLL %d not locked\n", id);
> >   }
> > -static void skl_ddi_dpll0_enable(struct drm_i915_private *dev_priv,
> > +static void skl_ddi_dpll0_enable(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll)
> >   {
> > -	skl_ddi_pll_write_ctrl1(dev_priv, pll);
> > +	skl_ddi_pll_write_ctrl1(i915, pll);
> >   }
> > -static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> > +static void skl_ddi_pll_disable(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll)
> >   {
> >   	const struct skl_dpll_regs *regs = skl_dpll_regs;
> > @@ -988,12 +988,12 @@ static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(regs[id].ctl);
> >   }
> > -static void skl_ddi_dpll0_disable(struct drm_i915_private *dev_priv,
> > +static void skl_ddi_dpll0_disable(struct drm_i915_private *i915,
> >   				  struct intel_shared_dpll *pll)
> >   {
> >   }
> > -static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> >   				     struct intel_shared_dpll *pll,
> >   				     struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -1003,7 +1003,7 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1025,12 +1025,12 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > -static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *i915,
> >   				       struct intel_shared_dpll *pll,
> >   				       struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -1040,7 +1040,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> >   	u32 val;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1058,7 +1058,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > @@ -1424,7 +1424,7 @@ skl_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static void skl_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void skl_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: "
> > @@ -1446,7 +1446,7 @@ static const struct intel_shared_dpll_funcs skl_ddi_dpll0_funcs = {
> >   	.get_hw_state = skl_ddi_dpll0_get_hw_state,
> >   };
> > -static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> > +static void bxt_ddi_pll_enable(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll)
> >   {
> >   	u32 temp;
> > @@ -1454,14 +1454,14 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	enum dpio_phy phy;
> >   	enum dpio_channel ch;
> > -	bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> > +	bxt_port_to_phy_channel(i915, port, &phy, &ch);
> >   	/* Non-SSC reference */
> >   	temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> >   	temp |= PORT_PLL_REF_SEL;
> >   	I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> >   		temp |= PORT_PLL_POWER_ENABLE;
> >   		I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> > @@ -1549,7 +1549,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   			200))
> >   		DRM_ERROR("PLL %d not locked\n", port);
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		temp = I915_READ(BXT_PORT_TX_DW5_LN0(phy, ch));
> >   		temp |= DCC_DELAY_RANGE_2;
> >   		I915_WRITE(BXT_PORT_TX_DW5_GRP(phy, ch), temp);
> > @@ -1566,7 +1566,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(BXT_PORT_PCS_DW12_GRP(phy, ch), temp);
> >   }
> > -static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> > +static void bxt_ddi_pll_disable(struct drm_i915_private *i915,
> >   					struct intel_shared_dpll *pll)
> >   {
> >   	enum port port = (enum port)pll->info->id; /* 1:1 port->PLL mapping */
> > @@ -1577,7 +1577,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> >   	POSTING_READ(BXT_PORT_PLL_ENABLE(port));
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		temp = I915_READ(BXT_PORT_PLL_ENABLE(port));
> >   		temp &= ~PORT_PLL_POWER_ENABLE;
> >   		I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp);
> > @@ -1588,7 +1588,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> >   					struct intel_shared_dpll *pll,
> >   					struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -1599,9 +1599,9 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	u32 val;
> >   	bool ret;
> > -	bxt_port_to_phy_channel(dev_priv, port, &phy, &ch);
> > +	bxt_port_to_phy_channel(i915, port, &phy, &ch);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1660,7 +1660,7 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > @@ -1832,7 +1832,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
> >   	     struct intel_encoder *encoder)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_shared_dpll *pll;
> >   	enum intel_dpll_id id;
> > @@ -1846,7 +1846,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
> >   	/* 1:1 mapping between ports and PLLs */
> >   	id = (enum intel_dpll_id) encoder->port;
> > -	pll = intel_get_shared_dpll_by_id(dev_priv, id);
> > +	pll = intel_get_shared_dpll_by_id(i915, id);
> >   	DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n",
> >   		      crtc->base.base.id, crtc->base.name, pll->info->name);
> > @@ -1856,7 +1856,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static void bxt_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void bxt_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: ebb0: 0x%x, ebb4: 0x%x,"
> > @@ -1887,7 +1887,7 @@ struct intel_dpll_mgr {
> >   	struct intel_shared_dpll *(*get_dpll)(struct intel_crtc_state *crtc_state,
> >   					      struct intel_encoder *encoder);
> > -	void (*dump_hw_state)(struct drm_i915_private *dev_priv,
> > +	void (*dump_hw_state)(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state);
> >   };
> > @@ -1946,7 +1946,7 @@ static const struct intel_dpll_mgr bxt_pll_mgr = {
> >   	.dump_hw_state = bxt_dump_hw_state,
> >   };
> > -static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> > +static void cnl_ddi_pll_enable(struct drm_i915_private *i915,
> >   			       struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -1958,7 +1958,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CNL_DPLL_ENABLE(id), val);
> >   	/* 2. Wait for DPLL power state enabled in DPLL_ENABLE. */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    CNL_DPLL_ENABLE(id),
> >   				    PLL_POWER_STATE,
> >   				    PLL_POWER_STATE,
> > @@ -1999,7 +1999,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CNL_DPLL_ENABLE(id), val);
> >   	/* 7. Wait for PLL lock status in DPLL_ENABLE. */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    CNL_DPLL_ENABLE(id),
> >   				    PLL_LOCK,
> >   				    PLL_LOCK,
> > @@ -2021,7 +2021,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv,
> >   	 */
> >   }
> > -static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> > +static void cnl_ddi_pll_disable(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll)
> >   {
> >   	const enum intel_dpll_id id = pll->info->id;
> > @@ -2047,7 +2047,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CNL_DPLL_ENABLE(id), val);
> >   	/* 4. Wait for PLL not locked status in DPLL_ENABLE. */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    CNL_DPLL_ENABLE(id),
> >   				    PLL_LOCK,
> >   				    0,
> > @@ -2069,7 +2069,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(CNL_DPLL_ENABLE(id), val);
> >   	/* 7. Wait for DPLL power state disabled in DPLL_ENABLE. */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    CNL_DPLL_ENABLE(id),
> >   				    PLL_POWER_STATE,
> >   				    0,
> > @@ -2077,7 +2077,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv,
> >   		DRM_ERROR("PLL %d Power not disabled\n", id);
> >   }
> > -static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *i915,
> >   				     struct intel_shared_dpll *pll,
> >   				     struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -2086,7 +2086,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	u32 val;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -2107,7 +2107,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > @@ -2199,15 +2199,15 @@ static void cnl_wrpll_params_populate(struct skl_wrpll_params *params,
> >   	params->dco_fraction = dco & 0x7fff;
> >   }
> > -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv)
> > +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915)
> >   {
> > -	int ref_clock = dev_priv->cdclk.hw.ref;
> > +	int ref_clock = i915->cdclk.hw.ref;
> >   	/*
> >   	 * For ICL+, the spec states: if reference frequency is 38.4,
> >   	 * use 19.2 because the DPLL automatically divides that by 2.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 11 && ref_clock == 38400)
> > +	if (INTEL_GEN(i915) >= 11 && ref_clock == 38400)
> >   		ref_clock = 19200;
> >   	return ref_clock;
> > @@ -2217,7 +2217,7 @@ static bool
> >   cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state,
> >   			struct skl_wrpll_params *wrpll_params)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	u32 afe_clock = crtc_state->port_clock * 5;
> >   	u32 ref_clock;
> >   	u32 dco_min = 7998000;
> > @@ -2252,7 +2252,7 @@ cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state,
> >   	cnl_wrpll_get_multipliers(best_div, &pdiv, &qdiv, &kdiv);
> > -	ref_clock = cnl_hdmi_pll_ref_clock(dev_priv);
> > +	ref_clock = cnl_hdmi_pll_ref_clock(i915);
> >   	cnl_wrpll_params_populate(wrpll_params, best_dco, ref_clock,
> >   				  pdiv, qdiv, kdiv);
> > @@ -2370,7 +2370,7 @@ cnl_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static void cnl_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void cnl_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: "
> > @@ -2476,9 +2476,9 @@ static const struct skl_wrpll_params icl_tbt_pll_19_2MHz_values = {
> >   static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state,
> >   				  struct skl_wrpll_params *pll_params)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	const struct icl_combo_pll_params *params =
> > -		dev_priv->cdclk.hw.ref == 24000 ?
> > +		i915->cdclk.hw.ref == 24000 ?
> >   		icl_dp_combo_pll_24MHz_values :
> >   		icl_dp_combo_pll_19_2MHz_values;
> >   	int clock = crtc_state->port_clock;
> > @@ -2498,9 +2498,9 @@ static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state,
> >   static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state,
> >   			     struct skl_wrpll_params *pll_params)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	*pll_params = dev_priv->cdclk.hw.ref == 24000 ?
> > +	*pll_params = i915->cdclk.hw.ref == 24000 ?
> >   			icl_tbt_pll_24MHz_values : icl_tbt_pll_19_2MHz_values;
> >   	return true;
> >   }
> > @@ -2508,12 +2508,12 @@ static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state,
> >   static bool icl_calc_dpll_state(struct intel_crtc_state *crtc_state,
> >   				struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	u32 cfgcr0, cfgcr1;
> >   	struct skl_wrpll_params pll_params = { 0 };
> >   	bool ret;
> > -	if (intel_port_is_tc(dev_priv, encoder->port))
> > +	if (intel_port_is_tc(i915, encoder->port))
> >   		ret = icl_calc_tbt_pll(crtc_state, &pll_params);
> >   	else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI) ||
> >   		 intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI))
> > @@ -2629,9 +2629,9 @@ static bool icl_mg_pll_find_divisors(int clock_khz, bool is_dp, bool use_ssc,
> >    */
> >   static bool icl_calc_mg_pll_state(struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_dpll_hw_state *pll_state = &crtc_state->dpll_hw_state;
> > -	int refclk_khz = dev_priv->cdclk.hw.ref;
> > +	int refclk_khz = i915->cdclk.hw.ref;
> >   	int clock = crtc_state->port_clock;
> >   	u32 dco_khz, m1div, m2div_int, m2div_rem, m2div_frac;
> >   	u32 iref_ndiv, iref_trim, iref_pulse_w;
> > @@ -2796,18 +2796,18 @@ static struct intel_shared_dpll *
> >   icl_get_dpll(struct intel_crtc_state *crtc_state,
> >   	     struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_digital_port *intel_dig_port;
> >   	struct intel_shared_dpll *pll;
> >   	enum port port = encoder->port;
> >   	enum intel_dpll_id min, max;
> >   	bool ret;
> > -	if (intel_port_is_combophy(dev_priv, port)) {
> > +	if (intel_port_is_combophy(i915, port)) {
> >   		min = DPLL_ID_ICL_DPLL0;
> >   		max = DPLL_ID_ICL_DPLL1;
> >   		ret = icl_calc_dpll_state(crtc_state, encoder);
> > -	} else if (intel_port_is_tc(dev_priv, port)) {
> > +	} else if (intel_port_is_tc(i915, port)) {
> >   		if (encoder->type == INTEL_OUTPUT_DP_MST) {
> >   			struct intel_dp_mst_encoder *mst_encoder;
> > @@ -2824,7 +2824,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state,
> >   		} else {
> >   			enum tc_port tc_port;
> > -			tc_port = intel_port_to_tc(dev_priv, port);
> > +			tc_port = intel_port_to_tc(i915, port);
> >   			min = icl_tc_port_to_pll_id(tc_port);
> >   			max = min;
> >   			ret = icl_calc_mg_pll_state(crtc_state);
> > @@ -2851,7 +2851,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state,
> >   	return pll;
> >   }
> > -static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool mg_pll_get_hw_state(struct drm_i915_private *i915,
> >   				struct intel_shared_dpll *pll,
> >   				struct intel_dpll_hw_state *hw_state)
> >   {
> > @@ -2861,7 +2861,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	bool ret = false;
> >   	u32 val;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -2896,7 +2896,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	hw_state->mg_pll_tdc_coldst_bias =
> >   		I915_READ(MG_PLL_TDC_COLDST_BIAS(tc_port));
> > -	if (dev_priv->cdclk.hw.ref == 38400) {
> > +	if (i915->cdclk.hw.ref == 38400) {
> >   		hw_state->mg_pll_tdc_coldst_bias_mask = MG_PLL_TDC_COLDST_COLDSTART;
> >   		hw_state->mg_pll_bias_mask = 0;
> >   	} else {
> > @@ -2909,11 +2909,11 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > -static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool icl_pll_get_hw_state(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll,
> >   				 struct intel_dpll_hw_state *hw_state,
> >   				 i915_reg_t enable_reg)
> > @@ -2923,7 +2923,7 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	bool ret = false;
> >   	u32 val;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     POWER_DOMAIN_DISPLAY_CORE);
> >   	if (!wakeref)
> >   		return false;
> > @@ -2937,26 +2937,26 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv,
> >   	ret = true;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref);
> >   	return ret;
> >   }
> > -static bool combo_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool combo_pll_get_hw_state(struct drm_i915_private *i915,
> >   				   struct intel_shared_dpll *pll,
> >   				   struct intel_dpll_hw_state *hw_state)
> >   {
> > -	return icl_pll_get_hw_state(dev_priv, pll, hw_state,
> > +	return icl_pll_get_hw_state(i915, pll, hw_state,
> >   				    CNL_DPLL_ENABLE(pll->info->id));
> >   }
> > -static bool tbt_pll_get_hw_state(struct drm_i915_private *dev_priv,
> > +static bool tbt_pll_get_hw_state(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll,
> >   				 struct intel_dpll_hw_state *hw_state)
> >   {
> > -	return icl_pll_get_hw_state(dev_priv, pll, hw_state, TBT_PLL_ENABLE);
> > +	return icl_pll_get_hw_state(i915, pll, hw_state, TBT_PLL_ENABLE);
> >   }
> > -static void icl_dpll_write(struct drm_i915_private *dev_priv,
> > +static void icl_dpll_write(struct drm_i915_private *i915,
> >   			   struct intel_shared_dpll *pll)
> >   {
> >   	struct intel_dpll_hw_state *hw_state = &pll->state.hw_state;
> > @@ -2967,7 +2967,7 @@ static void icl_dpll_write(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(ICL_DPLL_CFGCR1(id));
> >   }
> > -static void icl_mg_pll_write(struct drm_i915_private *dev_priv,
> > +static void icl_mg_pll_write(struct drm_i915_private *i915,
> >   			     struct intel_shared_dpll *pll)
> >   {
> >   	struct intel_dpll_hw_state *hw_state = &pll->state.hw_state;
> > @@ -3017,7 +3017,7 @@ static void icl_mg_pll_write(struct drm_i915_private *dev_priv,
> >   	POSTING_READ(MG_PLL_TDC_COLDST_BIAS(tc_port));
> >   }
> > -static void icl_pll_power_enable(struct drm_i915_private *dev_priv,
> > +static void icl_pll_power_enable(struct drm_i915_private *i915,
> >   				 struct intel_shared_dpll *pll,
> >   				 i915_reg_t enable_reg)
> >   {
> > @@ -3031,12 +3031,12 @@ static void icl_pll_power_enable(struct drm_i915_private *dev_priv,
> >   	 * The spec says we need to "wait" but it also says it should be
> >   	 * immediate.
> >   	 */
> > -	if (intel_wait_for_register(&dev_priv->uncore, enable_reg,
> > +	if (intel_wait_for_register(&i915->uncore, enable_reg,
> >   				    PLL_POWER_STATE, PLL_POWER_STATE, 1))
> >   		DRM_ERROR("PLL %d Power not enabled\n", pll->info->id);
> >   }
> > -static void icl_pll_enable(struct drm_i915_private *dev_priv,
> > +static void icl_pll_enable(struct drm_i915_private *i915,
> >   			   struct intel_shared_dpll *pll,
> >   			   i915_reg_t enable_reg)
> >   {
> > @@ -3047,19 +3047,19 @@ static void icl_pll_enable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(enable_reg, val);
> >   	/* Timeout is actually 600us. */
> > -	if (intel_wait_for_register(&dev_priv->uncore, enable_reg,
> > +	if (intel_wait_for_register(&i915->uncore, enable_reg,
> >   				    PLL_LOCK, PLL_LOCK, 1))
> >   		DRM_ERROR("PLL %d not locked\n", pll->info->id);
> >   }
> > -static void combo_pll_enable(struct drm_i915_private *dev_priv,
> > +static void combo_pll_enable(struct drm_i915_private *i915,
> >   			     struct intel_shared_dpll *pll)
> >   {
> >   	i915_reg_t enable_reg = CNL_DPLL_ENABLE(pll->info->id);
> > -	icl_pll_power_enable(dev_priv, pll, enable_reg);
> > +	icl_pll_power_enable(i915, pll, enable_reg);
> > -	icl_dpll_write(dev_priv, pll);
> > +	icl_dpll_write(i915, pll);
> >   	/*
> >   	 * DVFS pre sequence would be here, but in our driver the cdclk code
> > @@ -3067,17 +3067,17 @@ static void combo_pll_enable(struct drm_i915_private *dev_priv,
> >   	 * nothing here.
> >   	 */
> > -	icl_pll_enable(dev_priv, pll, enable_reg);
> > +	icl_pll_enable(i915, pll, enable_reg);
> >   	/* DVFS post sequence would be here. See the comment above. */
> >   }
> > -static void tbt_pll_enable(struct drm_i915_private *dev_priv,
> > +static void tbt_pll_enable(struct drm_i915_private *i915,
> >   			   struct intel_shared_dpll *pll)
> >   {
> > -	icl_pll_power_enable(dev_priv, pll, TBT_PLL_ENABLE);
> > +	icl_pll_power_enable(i915, pll, TBT_PLL_ENABLE);
> > -	icl_dpll_write(dev_priv, pll);
> > +	icl_dpll_write(i915, pll);
> >   	/*
> >   	 * DVFS pre sequence would be here, but in our driver the cdclk code
> > @@ -3085,20 +3085,20 @@ static void tbt_pll_enable(struct drm_i915_private *dev_priv,
> >   	 * nothing here.
> >   	 */
> > -	icl_pll_enable(dev_priv, pll, TBT_PLL_ENABLE);
> > +	icl_pll_enable(i915, pll, TBT_PLL_ENABLE);
> >   	/* DVFS post sequence would be here. See the comment above. */
> >   }
> > -static void mg_pll_enable(struct drm_i915_private *dev_priv,
> > +static void mg_pll_enable(struct drm_i915_private *i915,
> >   			  struct intel_shared_dpll *pll)
> >   {
> >   	i915_reg_t enable_reg =
> >   		MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id));
> > -	icl_pll_power_enable(dev_priv, pll, enable_reg);
> > +	icl_pll_power_enable(i915, pll, enable_reg);
> > -	icl_mg_pll_write(dev_priv, pll);
> > +	icl_mg_pll_write(i915, pll);
> >   	/*
> >   	 * DVFS pre sequence would be here, but in our driver the cdclk code
> > @@ -3106,12 +3106,12 @@ static void mg_pll_enable(struct drm_i915_private *dev_priv,
> >   	 * nothing here.
> >   	 */
> > -	icl_pll_enable(dev_priv, pll, enable_reg);
> > +	icl_pll_enable(i915, pll, enable_reg);
> >   	/* DVFS post sequence would be here. See the comment above. */
> >   }
> > -static void icl_pll_disable(struct drm_i915_private *dev_priv,
> > +static void icl_pll_disable(struct drm_i915_private *i915,
> >   			    struct intel_shared_dpll *pll,
> >   			    i915_reg_t enable_reg)
> >   {
> > @@ -3130,7 +3130,7 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(enable_reg, val);
> >   	/* Timeout is actually 1us. */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    enable_reg, PLL_LOCK, 0, 1))
> >   		DRM_ERROR("PLL %d locked\n", pll->info->id);
> > @@ -3144,33 +3144,33 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv,
> >   	 * The spec says we need to "wait" but it also says it should be
> >   	 * immediate.
> >   	 */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    enable_reg, PLL_POWER_STATE, 0, 1))
> >   		DRM_ERROR("PLL %d Power not disabled\n", pll->info->id);
> >   }
> > -static void combo_pll_disable(struct drm_i915_private *dev_priv,
> > +static void combo_pll_disable(struct drm_i915_private *i915,
> >   			      struct intel_shared_dpll *pll)
> >   {
> > -	icl_pll_disable(dev_priv, pll, CNL_DPLL_ENABLE(pll->info->id));
> > +	icl_pll_disable(i915, pll, CNL_DPLL_ENABLE(pll->info->id));
> >   }
> > -static void tbt_pll_disable(struct drm_i915_private *dev_priv,
> > +static void tbt_pll_disable(struct drm_i915_private *i915,
> >   			    struct intel_shared_dpll *pll)
> >   {
> > -	icl_pll_disable(dev_priv, pll, TBT_PLL_ENABLE);
> > +	icl_pll_disable(i915, pll, TBT_PLL_ENABLE);
> >   }
> > -static void mg_pll_disable(struct drm_i915_private *dev_priv,
> > +static void mg_pll_disable(struct drm_i915_private *i915,
> >   			   struct intel_shared_dpll *pll)
> >   {
> >   	i915_reg_t enable_reg =
> >   		MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id));
> > -	icl_pll_disable(dev_priv, pll, enable_reg);
> > +	icl_pll_disable(i915, pll, enable_reg);
> >   }
> > -static void icl_dump_hw_state(struct drm_i915_private *dev_priv,
> > +static void icl_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> >   	DRM_DEBUG_KMS("dpll_hw_state: cfgcr0: 0x%x, cfgcr1: 0x%x, "
> > @@ -3247,28 +3247,28 @@ static const struct intel_dpll_mgr ehl_pll_mgr = {
> >    */
> >   void intel_shared_dpll_init(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	const struct intel_dpll_mgr *dpll_mgr = NULL;
> >   	const struct dpll_info *dpll_info;
> >   	int i;
> > -	if (IS_ELKHARTLAKE(dev_priv))
> > +	if (IS_ELKHARTLAKE(i915))
> >   		dpll_mgr = &ehl_pll_mgr;
> > -	else if (INTEL_GEN(dev_priv) >= 11)
> > +	else if (INTEL_GEN(i915) >= 11)
> >   		dpll_mgr = &icl_pll_mgr;
> > -	else if (IS_CANNONLAKE(dev_priv))
> > +	else if (IS_CANNONLAKE(i915))
> >   		dpll_mgr = &cnl_pll_mgr;
> > -	else if (IS_GEN9_BC(dev_priv))
> > +	else if (IS_GEN9_BC(i915))
> >   		dpll_mgr = &skl_pll_mgr;
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		dpll_mgr = &bxt_pll_mgr;
> > -	else if (HAS_DDI(dev_priv))
> > +	else if (HAS_DDI(i915))
> >   		dpll_mgr = &hsw_pll_mgr;
> > -	else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))
> > +	else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))
> >   		dpll_mgr = &pch_pll_mgr;
> >   	if (!dpll_mgr) {
> > -		dev_priv->num_shared_dpll = 0;
> > +		i915->num_shared_dpll = 0;
> >   		return;
> >   	}
> > @@ -3276,14 +3276,14 @@ void intel_shared_dpll_init(struct drm_device *dev)
> >   	for (i = 0; dpll_info[i].name; i++) {
> >   		WARN_ON(i != dpll_info[i].id);
> > -		dev_priv->shared_dplls[i].info = &dpll_info[i];
> > +		i915->shared_dplls[i].info = &dpll_info[i];
> >   	}
> > -	dev_priv->dpll_mgr = dpll_mgr;
> > -	dev_priv->num_shared_dpll = i;
> > -	mutex_init(&dev_priv->dpll_lock);
> > +	i915->dpll_mgr = dpll_mgr;
> > +	i915->num_shared_dpll = i;
> > +	mutex_init(&i915->dpll_lock);
> > -	BUG_ON(dev_priv->num_shared_dpll > I915_NUM_PLLS);
> > +	BUG_ON(i915->num_shared_dpll > I915_NUM_PLLS);
> >   }
> >   /**
> > @@ -3304,8 +3304,8 @@ struct intel_shared_dpll *
> >   intel_get_shared_dpll(struct intel_crtc_state *crtc_state,
> >   		      struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > -	const struct intel_dpll_mgr *dpll_mgr = dev_priv->dpll_mgr;
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > +	const struct intel_dpll_mgr *dpll_mgr = i915->dpll_mgr;
> >   	if (WARN_ON(!dpll_mgr))
> >   		return NULL;
> > @@ -3335,16 +3335,16 @@ void intel_release_shared_dpll(struct intel_shared_dpll *dpll,
> >   /**
> >    * intel_shared_dpll_dump_hw_state - write hw_state to dmesg
> > - * @dev_priv: i915 drm device
> > + * @i915: i915 drm device
> >    * @hw_state: hw state to be written to the log
> >    *
> >    * Write the relevant values in @hw_state to dmesg using DRM_DEBUG_KMS.
> >    */
> > -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv,
> > +void intel_dpll_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state)
> >   {
> > -	if (dev_priv->dpll_mgr) {
> > -		dev_priv->dpll_mgr->dump_hw_state(dev_priv, hw_state);
> > +	if (i915->dpll_mgr) {
> > +		i915->dpll_mgr->dump_hw_state(i915, hw_state);
> >   	} else {
> >   		/* fallback for platforms that don't use the shared dpll
> >   		 * infrastructure
> > diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.h b/drivers/gpu/drm/i915/intel_dpll_mgr.h
> > index d0570414f3d1..b068c8441328 100644
> > --- a/drivers/gpu/drm/i915/intel_dpll_mgr.h
> > +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.h
> > @@ -221,7 +221,7 @@ struct intel_shared_dpll_funcs {
> >   	 * Called from intel_prepare_shared_dpll() function unless the PLL
> >   	 * is already enabled.
> >   	 */
> > -	void (*prepare)(struct drm_i915_private *dev_priv,
> > +	void (*prepare)(struct drm_i915_private *i915,
> >   			struct intel_shared_dpll *pll);
> >   	/**
> > @@ -230,7 +230,7 @@ struct intel_shared_dpll_funcs {
> >   	 * Hook for enabling the pll, called from intel_enable_shared_dpll()
> >   	 * if the pll is not already enabled.
> >   	 */
> > -	void (*enable)(struct drm_i915_private *dev_priv,
> > +	void (*enable)(struct drm_i915_private *i915,
> >   		       struct intel_shared_dpll *pll);
> >   	/**
> > @@ -240,7 +240,7 @@ struct intel_shared_dpll_funcs {
> >   	 * only when it is safe to disable the pll, i.e., there are no more
> >   	 * tracked users for it.
> >   	 */
> > -	void (*disable)(struct drm_i915_private *dev_priv,
> > +	void (*disable)(struct drm_i915_private *i915,
> >   			struct intel_shared_dpll *pll);
> >   	/**
> > @@ -250,7 +250,7 @@ struct intel_shared_dpll_funcs {
> >   	 * registers. This is used for initial hw state readout and state
> >   	 * verification after a mode set.
> >   	 */
> > -	bool (*get_hw_state)(struct drm_i915_private *dev_priv,
> > +	bool (*get_hw_state)(struct drm_i915_private *i915,
> >   			     struct intel_shared_dpll *pll,
> >   			     struct intel_dpll_hw_state *hw_state);
> >   };
> > @@ -271,7 +271,7 @@ struct dpll_info {
> >   	/**
> >   	 * @id: unique indentifier for this DPLL; should match the index in the
> > -	 * dev_priv->shared_dplls array
> > +	 * i915->shared_dplls array
> >   	 */
> >   	enum intel_dpll_id id;
> > @@ -321,12 +321,12 @@ struct intel_shared_dpll {
> >   /* shared dpll functions */
> >   struct intel_shared_dpll *
> > -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv,
> > +intel_get_shared_dpll_by_id(struct drm_i915_private *i915,
> >   			    enum intel_dpll_id id);
> >   enum intel_dpll_id
> > -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv,
> > +intel_get_shared_dpll_id(struct drm_i915_private *i915,
> >   			 struct intel_shared_dpll *pll);
> > -void assert_shared_dpll(struct drm_i915_private *dev_priv,
> > +void assert_shared_dpll(struct drm_i915_private *i915,
> >   			struct intel_shared_dpll *pll,
> >   			bool state);
> >   #define assert_shared_dpll_enabled(d, p) assert_shared_dpll(d, p, true)
> > @@ -342,9 +342,9 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state);
> >   void intel_shared_dpll_swap_state(struct drm_atomic_state *state);
> >   void intel_shared_dpll_init(struct drm_device *dev);
> > -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv,
> > +void intel_dpll_dump_hw_state(struct drm_i915_private *i915,
> >   			      const struct intel_dpll_hw_state *hw_state);
> > -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv);
> > +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915);
> >   enum intel_dpll_id icl_tc_port_to_pll_id(enum tc_port tc_port);
> >   bool intel_dpll_is_combophy(enum intel_dpll_id id);
> > diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
> > index 3e337317f77e..23065dda82d9 100644
> > --- a/drivers/gpu/drm/i915/intel_drv.h
> > +++ b/drivers/gpu/drm/i915/intel_drv.h
> > @@ -963,7 +963,7 @@ struct intel_crtc {
> >   	struct intel_crtc_state *config;
> > -	/* Access to these should be protected by dev_priv->irq_lock. */
> > +	/* Access to these should be protected by i915->irq_lock. */
> >   	bool cpu_fifo_underrun_disabled;
> >   	bool pch_fifo_underrun_disabled;
> > @@ -1293,15 +1293,15 @@ vlv_pipe_to_channel(enum pipe pipe)
> >   }
> >   static inline struct intel_crtc *
> > -intel_get_crtc_for_pipe(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +intel_get_crtc_for_pipe(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> > -	return dev_priv->pipe_to_crtc_mapping[pipe];
> > +	return i915->pipe_to_crtc_mapping[pipe];
> >   }
> >   static inline struct intel_crtc *
> > -intel_get_crtc_for_plane(struct drm_i915_private *dev_priv, enum i9xx_plane_id plane)
> > +intel_get_crtc_for_plane(struct drm_i915_private *i915, enum i9xx_plane_id plane)
> >   {
> > -	return dev_priv->plane_to_crtc_mapping[plane];
> > +	return i915->plane_to_crtc_mapping[plane];
> >   }
> >   struct intel_load_detect_pipe {
> > @@ -1447,17 +1447,17 @@ intel_atomic_get_new_crtc_state(struct intel_atomic_state *state,
> >   /* intel_display.c */
> >   void intel_plane_destroy(struct drm_plane *plane);
> > -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe);
> > -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe);
> > +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe);
> > +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe);
> >   enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc);
> > -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv);
> > -int vlv_get_cck_clock(struct drm_i915_private *dev_priv,
> > +int vlv_get_hpll_vco(struct drm_i915_private *i915);
> > +int vlv_get_cck_clock(struct drm_i915_private *i915,
> >   		      const char *name, u32 reg, int ref_freq);
> > -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv,
> > +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915,
> >   			   const char *name, u32 reg);
> > -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv);
> > -void lpt_disable_iclkip(struct drm_i915_private *dev_priv);
> > -void intel_init_display_hooks(struct drm_i915_private *dev_priv);
> > +void lpt_disable_pch_transcoder(struct drm_i915_private *i915);
> > +void lpt_disable_iclkip(struct drm_i915_private *i915);
> > +void intel_init_display_hooks(struct drm_i915_private *i915);
> >   unsigned int intel_fb_xy_to_linear(int x, int y,
> >   				   const struct intel_plane_state *state,
> >   				   int plane);
> > @@ -1467,19 +1467,19 @@ void intel_add_fb_offsets(int *x, int *y,
> >   			  const struct intel_plane_state *state, int plane);
> >   unsigned int intel_rotation_info_size(const struct intel_rotation_info *rot_info);
> >   unsigned int intel_remapped_info_size(const struct intel_remapped_info *rem_info);
> > -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv);
> > +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915);
> >   int intel_display_suspend(struct drm_device *dev);
> > -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv);
> > +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915);
> >   void intel_encoder_destroy(struct drm_encoder *encoder);
> >   struct drm_display_mode *
> >   intel_encoder_current_mode(struct intel_encoder *encoder);
> > -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port);
> > -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port);
> > -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv,
> > +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port);
> > +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port);
> > +enum tc_port intel_port_to_tc(struct drm_i915_private *i915,
> >   			      enum port port);
> >   int intel_get_pipe_from_crtc_id_ioctl(struct drm_device *dev, void *data,
> >   				      struct drm_file *file_priv);
> > -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv,
> > +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915,
> >   					     enum pipe pipe);
> >   static inline bool
> >   intel_crtc_has_type(const struct intel_crtc_state *crtc_state,
> > @@ -1496,23 +1496,23 @@ intel_crtc_has_dp_encoder(const struct intel_crtc_state *crtc_state)
> >   		 (1 << INTEL_OUTPUT_EDP));
> >   }
> >   static inline void
> > -intel_wait_for_vblank(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +intel_wait_for_vblank(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> > -	drm_wait_one_vblank(&dev_priv->drm, pipe);
> > +	drm_wait_one_vblank(&i915->drm, pipe);
> >   }
> >   static inline void
> > -intel_wait_for_vblank_if_active(struct drm_i915_private *dev_priv, int pipe)
> > +intel_wait_for_vblank_if_active(struct drm_i915_private *i915, int pipe)
> >   {
> > -	const struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	const struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	if (crtc->active)
> > -		intel_wait_for_vblank(dev_priv, pipe);
> > +		intel_wait_for_vblank(i915, pipe);
> >   }
> >   u32 intel_crtc_get_vblank_counter(struct intel_crtc *crtc);
> >   int ironlake_get_lanes_required(int target_clock, int link_bw, int bpp);
> > -void vlv_wait_port_ready(struct drm_i915_private *dev_priv,
> > +void vlv_wait_port_ready(struct drm_i915_private *i915,
> >   			 struct intel_digital_port *dport,
> >   			 unsigned int expected_mask);
> >   int intel_get_load_detect_pipe(struct drm_connector *connector,
> > @@ -1536,34 +1536,34 @@ int intel_prepare_plane_fb(struct drm_plane *plane,
> >   void intel_cleanup_plane_fb(struct drm_plane *plane,
> >   			    struct drm_plane_state *old_state);
> > -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv,
> > +void assert_pch_transcoder_disabled(struct drm_i915_private *i915,
> >   				    enum pipe pipe);
> > -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe,
> > +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe,
> >   		     const struct dpll *dpll);
> > -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe);
> > -int lpt_get_iclkip(struct drm_i915_private *dev_priv);
> > +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe);
> > +int lpt_get_iclkip(struct drm_i915_private *i915);
> >   bool intel_fuzzy_clock_check(int clock1, int clock2);
> >   /* modesetting asserts */
> > -void assert_panel_unlocked(struct drm_i915_private *dev_priv,
> > +void assert_panel_unlocked(struct drm_i915_private *i915,
> >   			   enum pipe pipe);
> > -void assert_pll(struct drm_i915_private *dev_priv,
> > +void assert_pll(struct drm_i915_private *i915,
> >   		enum pipe pipe, bool state);
> >   #define assert_pll_enabled(d, p) assert_pll(d, p, true)
> >   #define assert_pll_disabled(d, p) assert_pll(d, p, false)
> > -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state);
> > +void assert_dsi_pll(struct drm_i915_private *i915, bool state);
> >   #define assert_dsi_pll_enabled(d) assert_dsi_pll(d, true)
> >   #define assert_dsi_pll_disabled(d) assert_dsi_pll(d, false)
> > -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv,
> > +void assert_fdi_rx_pll(struct drm_i915_private *i915,
> >   		       enum pipe pipe, bool state);
> >   #define assert_fdi_rx_pll_enabled(d, p) assert_fdi_rx_pll(d, p, true)
> >   #define assert_fdi_rx_pll_disabled(d, p) assert_fdi_rx_pll(d, p, false)
> > -void assert_pipe(struct drm_i915_private *dev_priv, enum pipe pipe, bool state);
> > +void assert_pipe(struct drm_i915_private *i915, enum pipe pipe, bool state);
> >   #define assert_pipe_enabled(d, p) assert_pipe(d, p, true)
> >   #define assert_pipe_disabled(d, p) assert_pipe(d, p, false)
> > -void intel_prepare_reset(struct drm_i915_private *dev_priv);
> > -void intel_finish_reset(struct drm_i915_private *dev_priv);
> > +void intel_prepare_reset(struct drm_i915_private *i915);
> > +void intel_finish_reset(struct drm_i915_private *i915);
> >   void intel_dp_get_m_n(struct intel_crtc *crtc,
> >   		      struct intel_crtc_state *pipe_config);
> >   void intel_dp_set_m_n(const struct intel_crtc_state *crtc_state,
> > diff --git a/drivers/gpu/drm/i915/intel_dsi.c b/drivers/gpu/drm/i915/intel_dsi.c
> > index 5fec02aceaed..09414da3d818 100644
> > --- a/drivers/gpu/drm/i915/intel_dsi.c
> > +++ b/drivers/gpu/drm/i915/intel_dsi.c
> > @@ -113,14 +113,14 @@ struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi,
> >   enum drm_panel_orientation
> >   intel_dsi_get_panel_orientation(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum drm_panel_orientation orientation;
> > -	orientation = dev_priv->vbt.dsi.orientation;
> > +	orientation = i915->vbt.dsi.orientation;
> >   	if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> >   		return orientation;
> > -	orientation = dev_priv->vbt.orientation;
> > +	orientation = i915->vbt.orientation;
> >   	if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> >   		return orientation;
> > diff --git a/drivers/gpu/drm/i915/intel_dsi.h b/drivers/gpu/drm/i915/intel_dsi.h
> > index 6d20434636cd..2f477263d1f6 100644
> > --- a/drivers/gpu/drm/i915/intel_dsi.h
> > +++ b/drivers/gpu/drm/i915/intel_dsi.h
> > @@ -155,7 +155,7 @@ static inline u16 intel_dsi_encoder_ports(struct intel_encoder *encoder)
> >   }
> >   /* icl_dsi.c */
> > -void icl_dsi_init(struct drm_i915_private *dev_priv);
> > +void icl_dsi_init(struct drm_i915_private *i915);
> >   /* intel_dsi.c */
> >   int intel_dsi_bitrate(const struct intel_dsi *intel_dsi);
> > @@ -172,7 +172,7 @@ enum drm_mode_status intel_dsi_mode_valid(struct drm_connector *connector,
> >   struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi,
> >   					   const struct mipi_dsi_host_ops *funcs,
> >   					   enum port port);
> > -void vlv_dsi_init(struct drm_i915_private *dev_priv);
> > +void vlv_dsi_init(struct drm_i915_private *i915);
> >   /* vlv_dsi_pll.c */
> >   int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> > @@ -184,7 +184,7 @@ u32 vlv_dsi_get_pclk(struct intel_encoder *encoder,
> >   		     struct intel_crtc_state *config);
> >   void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port);
> > -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv);
> > +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915);
> >   int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> >   			struct intel_crtc_state *config);
> >   void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> > diff --git a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> > index 8c33262cb0b2..e5dd4962705c 100644
> > --- a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> > +++ b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c
> > @@ -159,11 +159,11 @@ static int dcs_setup_backlight(struct intel_connector *connector,
> >   int intel_dsi_dcs_init_backlight_funcs(struct intel_connector *intel_connector)
> >   {
> >   	struct drm_device *dev = intel_connector->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_encoder *encoder = intel_connector->encoder;
> >   	struct intel_panel *panel = &intel_connector->panel;
> > -	if (dev_priv->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS)
> > +	if (i915->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS)
> >   		return -ENODEV;
> >   	if (WARN_ON(encoder->type != INTEL_OUTPUT_DSI))
> > diff --git a/drivers/gpu/drm/i915/intel_dsi_vbt.c b/drivers/gpu/drm/i915/intel_dsi_vbt.c
> > index e5b178660408..d44ebdfd8760 100644
> > --- a/drivers/gpu/drm/i915/intel_dsi_vbt.c
> > +++ b/drivers/gpu/drm/i915/intel_dsi_vbt.c
> > @@ -121,7 +121,7 @@ static inline enum port intel_dsi_seq_port_to_port(u8 port)
> >   static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi,
> >   				       const u8 *data)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> >   	struct mipi_dsi_device *dsi_device;
> >   	u8 type, flags, seq_port;
> >   	u16 len;
> > @@ -192,7 +192,7 @@ static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi,
> >   		break;
> >   	}
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		vlv_dsi_wait_for_fifo_empty(intel_dsi, port);
> >   out:
> > @@ -213,7 +213,7 @@ static const u8 *mipi_exec_delay(struct intel_dsi *intel_dsi, const u8 *data)
> >   	return data;
> >   }
> > -static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
> > +static void vlv_exec_gpio(struct drm_i915_private *i915,
> >   			  u8 gpio_source, u8 gpio_index, bool value)
> >   {
> >   	struct gpio_map *map;
> > @@ -228,7 +228,7 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
> >   	map = &vlv_gpio_table[gpio_index];
> > -	if (dev_priv->vbt.dsi.seq_version >= 3) {
> > +	if (i915->vbt.dsi.seq_version >= 3) {
> >   		/* XXX: this assumes vlv_gpio_table only has NC GPIOs. */
> >   		port = IOSF_PORT_GPIO_NC;
> >   	} else {
> > @@ -246,26 +246,26 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv,
> >   	pconf0 = VLV_GPIO_PCONF0(map->base_offset);
> >   	padval = VLV_GPIO_PAD_VAL(map->base_offset);
> > -	vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> > +	vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO));
> >   	if (!map->init) {
> >   		/* FIXME: remove constant below */
> > -		vlv_iosf_sb_write(dev_priv, port, pconf0, 0x2000CC00);
> > +		vlv_iosf_sb_write(i915, port, pconf0, 0x2000CC00);
> >   		map->init = true;
> >   	}
> >   	tmp = 0x4 | value;
> > -	vlv_iosf_sb_write(dev_priv, port, padval, tmp);
> > -	vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> > +	vlv_iosf_sb_write(i915, port, padval, tmp);
> > +	vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO));
> >   }
> > -static void chv_exec_gpio(struct drm_i915_private *dev_priv,
> > +static void chv_exec_gpio(struct drm_i915_private *i915,
> >   			  u8 gpio_source, u8 gpio_index, bool value)
> >   {
> >   	u16 cfg0, cfg1;
> >   	u16 family_num;
> >   	u8 port;
> > -	if (dev_priv->vbt.dsi.seq_version >= 3) {
> > +	if (i915->vbt.dsi.seq_version >= 3) {
> >   		if (gpio_index >= CHV_GPIO_IDX_START_SE) {
> >   			/* XXX: it's unclear whether 255->57 is part of SE. */
> >   			gpio_index -= CHV_GPIO_IDX_START_SE;
> > @@ -301,15 +301,15 @@ static void chv_exec_gpio(struct drm_i915_private *dev_priv,
> >   	cfg0 = CHV_GPIO_PAD_CFG0(family_num, gpio_index);
> >   	cfg1 = CHV_GPIO_PAD_CFG1(family_num, gpio_index);
> > -	vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> > -	vlv_iosf_sb_write(dev_priv, port, cfg1, 0);
> > -	vlv_iosf_sb_write(dev_priv, port, cfg0,
> > +	vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO));
> > +	vlv_iosf_sb_write(i915, port, cfg1, 0);
> > +	vlv_iosf_sb_write(i915, port, cfg0,
> >   			  CHV_GPIO_GPIOEN | CHV_GPIO_GPIOCFG_GPO |
> >   			  CHV_GPIO_GPIOTXSTATE(value));
> > -	vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO));
> > +	vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO));
> >   }
> > -static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> > +static void bxt_exec_gpio(struct drm_i915_private *i915,
> >   			  u8 gpio_source, u8 gpio_index, bool value)
> >   {
> >   	/* XXX: this table is a quick ugly hack. */
> > @@ -317,7 +317,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> >   	struct gpio_desc *gpio_desc = bxt_gpio_table[gpio_index];
> >   	if (!gpio_desc) {
> > -		gpio_desc = devm_gpiod_get_index(dev_priv->drm.dev,
> > +		gpio_desc = devm_gpiod_get_index(i915->drm.dev,
> >   						 NULL, gpio_index,
> >   						 value ? GPIOD_OUT_LOW :
> >   						 GPIOD_OUT_HIGH);
> > @@ -334,7 +334,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv,
> >   	gpiod_set_value(gpio_desc, value);
> >   }
> > -static void icl_exec_gpio(struct drm_i915_private *dev_priv,
> > +static void icl_exec_gpio(struct drm_i915_private *i915,
> >   			  u8 gpio_source, u8 gpio_index, bool value)
> >   {
> >   	DRM_DEBUG_KMS("Skipping ICL GPIO element execution\n");
> > @@ -343,19 +343,19 @@ static void icl_exec_gpio(struct drm_i915_private *dev_priv,
> >   static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data)
> >   {
> >   	struct drm_device *dev = intel_dsi->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u8 gpio_source, gpio_index = 0, gpio_number;
> >   	bool value;
> >   	DRM_DEBUG_KMS("\n");
> > -	if (dev_priv->vbt.dsi.seq_version >= 3)
> > +	if (i915->vbt.dsi.seq_version >= 3)
> >   		gpio_index = *data++;
> >   	gpio_number = *data++;
> >   	/* gpio source in sequence v2 only */
> > -	if (dev_priv->vbt.dsi.seq_version == 2)
> > +	if (i915->vbt.dsi.seq_version == 2)
> >   		gpio_source = (*data >> 1) & 3;
> >   	else
> >   		gpio_source = 0;
> > @@ -363,14 +363,14 @@ static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data)
> >   	/* pull up/down */
> >   	value = *data++ & 1;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		icl_exec_gpio(dev_priv, gpio_source, gpio_index, value);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		vlv_exec_gpio(dev_priv, gpio_source, gpio_number, value);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		chv_exec_gpio(dev_priv, gpio_source, gpio_number, value);
> > +	if (INTEL_GEN(i915) >= 11)
> > +		icl_exec_gpio(i915, gpio_source, gpio_index, value);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		vlv_exec_gpio(i915, gpio_source, gpio_number, value);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		chv_exec_gpio(i915, gpio_source, gpio_number, value);
> >   	else
> > -		bxt_exec_gpio(dev_priv, gpio_source, gpio_index, value);
> > +		bxt_exec_gpio(i915, gpio_source, gpio_index, value);
> >   	return data;
> >   }
> > @@ -456,14 +456,14 @@ static const char *sequence_name(enum mipi_seq seq_id)
> >   void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> >   				 enum mipi_seq seq_id)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> >   	const u8 *data;
> >   	fn_mipi_elem_exec mipi_elem_exec;
> > -	if (WARN_ON(seq_id >= ARRAY_SIZE(dev_priv->vbt.dsi.sequence)))
> > +	if (WARN_ON(seq_id >= ARRAY_SIZE(i915->vbt.dsi.sequence)))
> >   		return;
> > -	data = dev_priv->vbt.dsi.sequence[seq_id];
> > +	data = i915->vbt.dsi.sequence[seq_id];
> >   	if (!data)
> >   		return;
> > @@ -476,7 +476,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> >   	data++;
> >   	/* Skip Size of Sequence. */
> > -	if (dev_priv->vbt.dsi.seq_version >= 3)
> > +	if (i915->vbt.dsi.seq_version >= 3)
> >   		data += 4;
> >   	while (1) {
> > @@ -492,7 +492,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> >   			mipi_elem_exec = NULL;
> >   		/* Size of Operation. */
> > -		if (dev_priv->vbt.dsi.seq_version >= 3)
> > +		if (i915->vbt.dsi.seq_version >= 3)
> >   			operation_size = *data++;
> >   		if (mipi_elem_exec) {
> > @@ -521,10 +521,10 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi,
> >   void intel_dsi_msleep(struct intel_dsi *intel_dsi, int msec)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev);
> >   	/* For v3 VBTs in vid-mode the delays are part of the VBT sequences */
> > -	if (is_vid_mode(intel_dsi) && dev_priv->vbt.dsi.seq_version >= 3)
> > +	if (is_vid_mode(intel_dsi) && i915->vbt.dsi.seq_version >= 3)
> >   		return;
> >   	msleep(msec);
> > @@ -571,10 +571,10 @@ void intel_dsi_log_params(struct intel_dsi *intel_dsi)
> >   bool intel_dsi_vbt_init(struct intel_dsi *intel_dsi, u16 panel_id)
> >   {
> >   	struct drm_device *dev = intel_dsi->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> > -	struct mipi_pps_data *pps = dev_priv->vbt.dsi.pps;
> > -	struct drm_display_mode *mode = dev_priv->vbt.lfp_lvds_vbt_mode;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct mipi_config *mipi_config = i915->vbt.dsi.config;
> > +	struct mipi_pps_data *pps = i915->vbt.dsi.pps;
> > +	struct drm_display_mode *mode = i915->vbt.lfp_lvds_vbt_mode;
> >   	u16 burst_mode_ratio;
> >   	enum port port;
> > diff --git a/drivers/gpu/drm/i915/intel_dvo.c b/drivers/gpu/drm/i915/intel_dvo.c
> > index 22666d28f4aa..6ced95b73538 100644
> > --- a/drivers/gpu/drm/i915/intel_dvo.c
> > +++ b/drivers/gpu/drm/i915/intel_dvo.c
> > @@ -133,7 +133,7 @@ static struct intel_dvo *intel_attached_dvo(struct drm_connector *connector)
> >   static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector)
> >   {
> >   	struct drm_device *dev = connector->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_dvo *intel_dvo = intel_attached_dvo(&connector->base);
> >   	u32 tmp;
> > @@ -148,7 +148,7 @@ static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector)
> >   static bool intel_dvo_get_hw_state(struct intel_encoder *encoder,
> >   				   enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> >   	u32 tmp;
> > @@ -162,7 +162,7 @@ static bool intel_dvo_get_hw_state(struct intel_encoder *encoder,
> >   static void intel_dvo_get_config(struct intel_encoder *encoder,
> >   				 struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> >   	u32 tmp, flags = 0;
> > @@ -187,7 +187,7 @@ static void intel_disable_dvo(struct intel_encoder *encoder,
> >   			      const struct intel_crtc_state *old_crtc_state,
> >   			      const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> >   	i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg;
> >   	u32 temp = I915_READ(dvo_reg);
> > @@ -201,7 +201,7 @@ static void intel_enable_dvo(struct intel_encoder *encoder,
> >   			     const struct intel_crtc_state *pipe_config,
> >   			     const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> >   	i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg;
> >   	u32 temp = I915_READ(dvo_reg);
> > @@ -276,7 +276,7 @@ static void intel_dvo_pre_enable(struct intel_encoder *encoder,
> >   				 const struct intel_crtc_state *pipe_config,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> >   	struct intel_dvo *intel_dvo = enc_to_dvo(encoder);
> > @@ -319,7 +319,7 @@ intel_dvo_detect(struct drm_connector *connector, bool force)
> >   static int intel_dvo_get_modes(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	const struct drm_display_mode *fixed_mode =
> >   		to_intel_connector(connector)->panel.fixed_mode;
> > @@ -330,7 +330,7 @@ static int intel_dvo_get_modes(struct drm_connector *connector)
> >   	 * that's not the case.
> >   	 */
> >   	intel_ddc_get_modes(connector,
> > -			    intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPC));
> > +			    intel_gmbus_get_adapter(i915, GMBUS_PIN_DPC));
> >   	if (!list_empty(&connector->probed_modes))
> >   		return 1;
> > @@ -406,7 +406,7 @@ static enum port intel_dvo_port(i915_reg_t dvo_reg)
> >   		return PORT_C;
> >   }
> > -void intel_dvo_init(struct drm_i915_private *dev_priv)
> > +void intel_dvo_init(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *intel_encoder;
> >   	struct intel_dvo *intel_dvo;
> > @@ -452,7 +452,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   		 * special cases, but otherwise default to what's defined
> >   		 * in the spec.
> >   		 */
> > -		if (intel_gmbus_is_valid_pin(dev_priv, dvo->gpio))
> > +		if (intel_gmbus_is_valid_pin(i915, dvo->gpio))
> >   			gpio = dvo->gpio;
> >   		else if (dvo->type == INTEL_DVO_CHIP_LVDS)
> >   			gpio = GMBUS_PIN_SSC;
> > @@ -464,7 +464,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   		 * It appears that everything is on GPIOE except for panels
> >   		 * on i830 laptops, which are on GPIOB (DVOA).
> >   		 */
> > -		i2c = intel_gmbus_get_adapter(dev_priv, gpio);
> > +		i2c = intel_gmbus_get_adapter(i915, gpio);
> >   		intel_dvo->dev = *dvo;
> > @@ -480,7 +480,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   		 * have the clock enabled before we attempt to
> >   		 * initialize the device.
> >   		 */
> > -		for_each_pipe(dev_priv, pipe) {
> > +		for_each_pipe(i915, pipe) {
> >   			dpll[pipe] = I915_READ(DPLL(pipe));
> >   			I915_WRITE(DPLL(pipe), dpll[pipe] | DPLL_DVO_2X_MODE);
> >   		}
> > @@ -488,7 +488,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   		dvoinit = dvo->dev_ops->init(&intel_dvo->dev, i2c);
> >   		/* restore the DVO 2x clock state to original */
> > -		for_each_pipe(dev_priv, pipe) {
> > +		for_each_pipe(i915, pipe) {
> >   			I915_WRITE(DPLL(pipe), dpll[pipe]);
> >   		}
> > @@ -498,7 +498,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   			continue;
> >   		port = intel_dvo_port(dvo->dvo_reg);
> > -		drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> > +		drm_encoder_init(&i915->drm, &intel_encoder->base,
> >   				 &intel_dvo_enc_funcs, encoder_type,
> >   				 "DVO %c", port_name(port));
> > @@ -511,14 +511,14 @@ void intel_dvo_init(struct drm_i915_private *dev_priv)
> >   		case INTEL_DVO_CHIP_TMDS:
> >   			intel_encoder->cloneable = (1 << INTEL_OUTPUT_ANALOG) |
> >   				(1 << INTEL_OUTPUT_DVO);
> > -			drm_connector_init(&dev_priv->drm, connector,
> > +			drm_connector_init(&i915->drm, connector,
> >   					   &intel_dvo_connector_funcs,
> >   					   DRM_MODE_CONNECTOR_DVII);
> >   			encoder_type = DRM_MODE_ENCODER_TMDS;
> >   			break;
> >   		case INTEL_DVO_CHIP_LVDS:
> >   			intel_encoder->cloneable = 0;
> > -			drm_connector_init(&dev_priv->drm, connector,
> > +			drm_connector_init(&i915->drm, connector,
> >   					   &intel_dvo_connector_funcs,
> >   					   DRM_MODE_CONNECTOR_LVDS);
> >   			encoder_type = DRM_MODE_ENCODER_LVDS;
> > diff --git a/drivers/gpu/drm/i915/intel_dvo.h b/drivers/gpu/drm/i915/intel_dvo.h
> > index 3ed0fdf8efff..d865cd439bb7 100644
> > --- a/drivers/gpu/drm/i915/intel_dvo.h
> > +++ b/drivers/gpu/drm/i915/intel_dvo.h
> > @@ -8,6 +8,6 @@
> >   struct drm_i915_private;
> > -void intel_dvo_init(struct drm_i915_private *dev_priv);
> > +void intel_dvo_init(struct drm_i915_private *i915);
> >   #endif /* __INTEL_DVO_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_dvo_dev.h b/drivers/gpu/drm/i915/intel_dvo_dev.h
> > index 94a6ae1e0292..1d184dc7922b 100644
> > --- a/drivers/gpu/drm/i915/intel_dvo_dev.h
> > +++ b/drivers/gpu/drm/i915/intel_dvo_dev.h
> > @@ -40,7 +40,7 @@ struct intel_dvo_device {
> >   	int slave_addr;
> >   	const struct intel_dvo_dev_ops *dev_ops;
> > -	void *dev_priv;
> > +	void *i915;
> >   	struct i2c_adapter *i2c_bus;
> >   };
> > diff --git a/drivers/gpu/drm/i915/intel_fbc.c b/drivers/gpu/drm/i915/intel_fbc.c
> > index 5679f2fffb7c..627fed2a55d2 100644
> > --- a/drivers/gpu/drm/i915/intel_fbc.c
> > +++ b/drivers/gpu/drm/i915/intel_fbc.c
> > @@ -45,14 +45,14 @@
> >   #include "intel_fbc.h"
> >   #include "intel_frontbuffer.h"
> > -static inline bool fbc_supported(struct drm_i915_private *dev_priv)
> > +static inline bool fbc_supported(struct drm_i915_private *i915)
> >   {
> > -	return HAS_FBC(dev_priv);
> > +	return HAS_FBC(i915);
> >   }
> > -static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *dev_priv)
> > +static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *i915)
> >   {
> > -	return INTEL_GEN(dev_priv) <= 3;
> > +	return INTEL_GEN(i915) <= 3;
> >   }
> >   /*
> > @@ -82,22 +82,22 @@ static void intel_fbc_get_plane_source_size(struct intel_fbc_state_cache *cache,
> >   		*height = cache->plane.src_h;
> >   }
> > -static int intel_fbc_calculate_cfb_size(struct drm_i915_private *dev_priv,
> > +static int intel_fbc_calculate_cfb_size(struct drm_i915_private *i915,
> >   					struct intel_fbc_state_cache *cache)
> >   {
> >   	int lines;
> >   	intel_fbc_get_plane_source_size(cache, NULL, &lines);
> > -	if (IS_GEN(dev_priv, 7))
> > +	if (IS_GEN(i915, 7))
> >   		lines = min(lines, 2048);
> > -	else if (INTEL_GEN(dev_priv) >= 8)
> > +	else if (INTEL_GEN(i915) >= 8)
> >   		lines = min(lines, 2560);
> >   	/* Hardware needs the full buffer stride, not just the active area. */
> >   	return lines * cache->fb.stride;
> >   }
> > -static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> > +static void i8xx_fbc_deactivate(struct drm_i915_private *i915)
> >   {
> >   	u32 fbc_ctl;
> > @@ -110,7 +110,7 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(FBC_CONTROL, fbc_ctl);
> >   	/* Wait for compressing bit to clear */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    FBC_STATUS, FBC_STAT_COMPRESSING, 0,
> >   				    10)) {
> >   		DRM_DEBUG_KMS("FBC idle timed out\n");
> > @@ -118,9 +118,9 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> > +static void i8xx_fbc_activate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> > +	struct intel_fbc_reg_params *params = &i915->fbc.params;
> >   	int cfb_pitch;
> >   	int i;
> >   	u32 fbc_ctl;
> > @@ -131,7 +131,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> >   		cfb_pitch = params->fb.stride;
> >   	/* FBC_CTL wants 32B or 64B units */
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		cfb_pitch = (cfb_pitch / 32) - 1;
> >   	else
> >   		cfb_pitch = (cfb_pitch / 64) - 1;
> > @@ -140,7 +140,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> >   	for (i = 0; i < (FBC_LL_SIZE / 32) + 1; i++)
> >   		I915_WRITE(FBC_TAG(i), 0);
> > -	if (IS_GEN(dev_priv, 4)) {
> > +	if (IS_GEN(i915, 4)) {
> >   		u32 fbc_ctl2;
> >   		/* Set it up... */
> > @@ -154,21 +154,21 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv)
> >   	fbc_ctl = I915_READ(FBC_CONTROL);
> >   	fbc_ctl &= 0x3fff << FBC_CTL_INTERVAL_SHIFT;
> >   	fbc_ctl |= FBC_CTL_EN | FBC_CTL_PERIODIC;
> > -	if (IS_I945GM(dev_priv))
> > +	if (IS_I945GM(i915))
> >   		fbc_ctl |= FBC_CTL_C3_IDLE; /* 945 needs special SR handling */
> >   	fbc_ctl |= (cfb_pitch & 0xff) << FBC_CTL_STRIDE_SHIFT;
> >   	fbc_ctl |= params->vma->fence->id;
> >   	I915_WRITE(FBC_CONTROL, fbc_ctl);
> >   }
> > -static bool i8xx_fbc_is_active(struct drm_i915_private *dev_priv)
> > +static bool i8xx_fbc_is_active(struct drm_i915_private *i915)
> >   {
> >   	return I915_READ(FBC_CONTROL) & FBC_CTL_EN;
> >   }
> > -static void g4x_fbc_activate(struct drm_i915_private *dev_priv)
> > +static void g4x_fbc_activate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> > +	struct intel_fbc_reg_params *params = &i915->fbc.params;
> >   	u32 dpfc_ctl;
> >   	dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane) | DPFC_SR_EN;
> > @@ -188,7 +188,7 @@ static void g4x_fbc_activate(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
> >   }
> > -static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv)
> > +static void g4x_fbc_deactivate(struct drm_i915_private *i915)
> >   {
> >   	u32 dpfc_ctl;
> > @@ -200,23 +200,23 @@ static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static bool g4x_fbc_is_active(struct drm_i915_private *dev_priv)
> > +static bool g4x_fbc_is_active(struct drm_i915_private *i915)
> >   {
> >   	return I915_READ(DPFC_CONTROL) & DPFC_CTL_EN;
> >   }
> >   /* This function forces a CFB recompression through the nuke operation. */
> > -static void intel_fbc_recompress(struct drm_i915_private *dev_priv)
> > +static void intel_fbc_recompress(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(MSG_FBC_REND_STATE, FBC_REND_NUKE);
> >   	POSTING_READ(MSG_FBC_REND_STATE);
> >   }
> > -static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> > +static void ilk_fbc_activate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> > +	struct intel_fbc_reg_params *params = &i915->fbc.params;
> >   	u32 dpfc_ctl;
> > -	int threshold = dev_priv->fbc.threshold;
> > +	int threshold = i915->fbc.threshold;
> >   	dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane);
> >   	if (params->fb.format->cpp[0] == 2)
> > @@ -237,9 +237,9 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> >   	if (params->flags & PLANE_HAS_FENCE) {
> >   		dpfc_ctl |= DPFC_CTL_FENCE_EN;
> > -		if (IS_GEN(dev_priv, 5))
> > +		if (IS_GEN(i915, 5))
> >   			dpfc_ctl |= params->vma->fence->id;
> > -		if (IS_GEN(dev_priv, 6)) {
> > +		if (IS_GEN(i915, 6)) {
> >   			I915_WRITE(SNB_DPFC_CTL_SA,
> >   				   SNB_CPU_FENCE_ENABLE |
> >   				   params->vma->fence->id);
> > @@ -247,7 +247,7 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> >   				   params->crtc.fence_y_offset);
> >   		}
> >   	} else {
> > -		if (IS_GEN(dev_priv, 6)) {
> > +		if (IS_GEN(i915, 6)) {
> >   			I915_WRITE(SNB_DPFC_CTL_SA, 0);
> >   			I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0);
> >   		}
> > @@ -259,10 +259,10 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv)
> >   	/* enable it... */
> >   	I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
> > -	intel_fbc_recompress(dev_priv);
> > +	intel_fbc_recompress(i915);
> >   }
> > -static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv)
> > +static void ilk_fbc_deactivate(struct drm_i915_private *i915)
> >   {
> >   	u32 dpfc_ctl;
> > @@ -274,19 +274,19 @@ static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static bool ilk_fbc_is_active(struct drm_i915_private *dev_priv)
> > +static bool ilk_fbc_is_active(struct drm_i915_private *i915)
> >   {
> >   	return I915_READ(ILK_DPFC_CONTROL) & DPFC_CTL_EN;
> >   }
> > -static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> > +static void gen7_fbc_activate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc_reg_params *params = &dev_priv->fbc.params;
> > +	struct intel_fbc_reg_params *params = &i915->fbc.params;
> >   	u32 dpfc_ctl;
> > -	int threshold = dev_priv->fbc.threshold;
> > +	int threshold = i915->fbc.threshold;
> >   	/* Display WA #0529: skl, kbl, bxt. */
> > -	if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915)) {
> >   		u32 val = I915_READ(CHICKEN_MISC_4);
> >   		val &= ~(FBC_STRIDE_OVERRIDE | FBC_STRIDE_MASK);
> > @@ -299,7 +299,7 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> >   	}
> >   	dpfc_ctl = 0;
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		dpfc_ctl |= IVB_DPFC_CTL_PLANE(params->crtc.i9xx_plane);
> >   	if (params->fb.format->cpp[0] == 2)
> > @@ -329,15 +329,15 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0);
> >   	}
> > -	if (dev_priv->fbc.false_color)
> > +	if (i915->fbc.false_color)
> >   		dpfc_ctl |= FBC_CTL_FALSE_COLOR;
> > -	if (IS_IVYBRIDGE(dev_priv)) {
> > +	if (IS_IVYBRIDGE(i915)) {
> >   		/* WaFbcAsynchFlipDisableFbcQueue:ivb */
> >   		I915_WRITE(ILK_DISPLAY_CHICKEN1,
> >   			   I915_READ(ILK_DISPLAY_CHICKEN1) |
> >   			   ILK_FBCQ_DIS);
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		/* WaFbcAsynchFlipDisableFbcQueue:hsw,bdw */
> >   		I915_WRITE(CHICKEN_PIPESL_1(params->crtc.pipe),
> >   			   I915_READ(CHICKEN_PIPESL_1(params->crtc.pipe)) |
> > @@ -346,72 +346,72 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN);
> > -	intel_fbc_recompress(dev_priv);
> > +	intel_fbc_recompress(i915);
> >   }
> > -static bool intel_fbc_hw_is_active(struct drm_i915_private *dev_priv)
> > +static bool intel_fbc_hw_is_active(struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 5)
> > -		return ilk_fbc_is_active(dev_priv);
> > -	else if (IS_GM45(dev_priv))
> > -		return g4x_fbc_is_active(dev_priv);
> > +	if (INTEL_GEN(i915) >= 5)
> > +		return ilk_fbc_is_active(i915);
> > +	else if (IS_GM45(i915))
> > +		return g4x_fbc_is_active(i915);
> >   	else
> > -		return i8xx_fbc_is_active(dev_priv);
> > +		return i8xx_fbc_is_active(i915);
> >   }
> > -static void intel_fbc_hw_activate(struct drm_i915_private *dev_priv)
> > +static void intel_fbc_hw_activate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	fbc->active = true;
> > -	if (INTEL_GEN(dev_priv) >= 7)
> > -		gen7_fbc_activate(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 5)
> > -		ilk_fbc_activate(dev_priv);
> > -	else if (IS_GM45(dev_priv))
> > -		g4x_fbc_activate(dev_priv);
> > +	if (INTEL_GEN(i915) >= 7)
> > +		gen7_fbc_activate(i915);
> > +	else if (INTEL_GEN(i915) >= 5)
> > +		ilk_fbc_activate(i915);
> > +	else if (IS_GM45(i915))
> > +		g4x_fbc_activate(i915);
> >   	else
> > -		i8xx_fbc_activate(dev_priv);
> > +		i8xx_fbc_activate(i915);
> >   }
> > -static void intel_fbc_hw_deactivate(struct drm_i915_private *dev_priv)
> > +static void intel_fbc_hw_deactivate(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	fbc->active = false;
> > -	if (INTEL_GEN(dev_priv) >= 5)
> > -		ilk_fbc_deactivate(dev_priv);
> > -	else if (IS_GM45(dev_priv))
> > -		g4x_fbc_deactivate(dev_priv);
> > +	if (INTEL_GEN(i915) >= 5)
> > +		ilk_fbc_deactivate(i915);
> > +	else if (IS_GM45(i915))
> > +		g4x_fbc_deactivate(i915);
> >   	else
> > -		i8xx_fbc_deactivate(dev_priv);
> > +		i8xx_fbc_deactivate(i915);
> >   }
> >   /**
> >    * intel_fbc_is_active - Is FBC active?
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function is used to verify the current state of FBC.
> >    *
> >    * FIXME: This should be tracked in the plane config eventually
> >    * instead of queried at runtime for most callers.
> >    */
> > -bool intel_fbc_is_active(struct drm_i915_private *dev_priv)
> > +bool intel_fbc_is_active(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->fbc.active;
> > +	return i915->fbc.active;
> >   }
> > -static void intel_fbc_deactivate(struct drm_i915_private *dev_priv,
> > +static void intel_fbc_deactivate(struct drm_i915_private *i915,
> >   				 const char *reason)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	WARN_ON(!mutex_is_locked(&fbc->lock));
> >   	if (fbc->active)
> > -		intel_fbc_hw_deactivate(dev_priv);
> > +		intel_fbc_hw_deactivate(i915);
> >   	fbc->no_fbc_reason = reason;
> >   }
> > @@ -419,12 +419,12 @@ static void intel_fbc_deactivate(struct drm_i915_private *dev_priv,
> >   static bool multiple_pipes_ok(struct intel_crtc *crtc,
> >   			      struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	enum pipe pipe = crtc->pipe;
> >   	/* Don't even bother tracking anything we don't need. */
> > -	if (!no_fbc_on_multiple_pipes(dev_priv))
> > +	if (!no_fbc_on_multiple_pipes(i915))
> >   		return true;
> >   	if (plane_state->base.visible)
> > @@ -435,7 +435,7 @@ static bool multiple_pipes_ok(struct intel_crtc *crtc,
> >   	return (fbc->visible_pipes_mask & ~(1 << pipe)) != 0;
> >   }
> > -static int find_compression_threshold(struct drm_i915_private *dev_priv,
> > +static int find_compression_threshold(struct drm_i915_private *i915,
> >   				      struct drm_mm_node *node,
> >   				      int size,
> >   				      int fb_cpp)
> > @@ -448,8 +448,8 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> >   	 * reserved range size, so it always assumes the maximum (8mb) is used.
> >   	 * If we enable FBC using a CFB on that memory range we'll get FIFO
> >   	 * underruns, even if that range is not reserved by the BIOS. */
> > -	if (IS_BROADWELL(dev_priv) || IS_GEN9_BC(dev_priv))
> > -		end = resource_size(&dev_priv->dsm) - 8 * 1024 * 1024;
> > +	if (IS_BROADWELL(i915) || IS_GEN9_BC(i915))
> > +		end = resource_size(&i915->dsm) - 8 * 1024 * 1024;
> >   	else
> >   		end = U64_MAX;
> > @@ -461,7 +461,7 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> >   	 */
> >   	/* Try to over-allocate to reduce reallocations and fragmentation. */
> > -	ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size <<= 1,
> > +	ret = i915_gem_stolen_insert_node_in_range(i915, node, size <<= 1,
> >   						   4096, 0, end);
> >   	if (ret == 0)
> >   		return compression_threshold;
> > @@ -472,9 +472,9 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> >   	    (fb_cpp == 2 && compression_threshold == 2))
> >   		return 0;
> > -	ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size >>= 1,
> > +	ret = i915_gem_stolen_insert_node_in_range(i915, node, size >>= 1,
> >   						   4096, 0, end);
> > -	if (ret && INTEL_GEN(dev_priv) <= 4) {
> > +	if (ret && INTEL_GEN(i915) <= 4) {
> >   		return 0;
> >   	} else if (ret) {
> >   		compression_threshold <<= 1;
> > @@ -486,17 +486,17 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv,
> >   static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct drm_mm_node *uninitialized_var(compressed_llb);
> >   	int size, fb_cpp, ret;
> >   	WARN_ON(drm_mm_node_allocated(&fbc->compressed_fb));
> > -	size = intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache);
> > +	size = intel_fbc_calculate_cfb_size(i915, &fbc->state_cache);
> >   	fb_cpp = fbc->state_cache.fb.format->cpp[0];
> > -	ret = find_compression_threshold(dev_priv, &fbc->compressed_fb,
> > +	ret = find_compression_threshold(i915, &fbc->compressed_fb,
> >   					 size, fb_cpp);
> >   	if (!ret)
> >   		goto err_llb;
> > @@ -507,32 +507,32 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
> >   	fbc->threshold = ret;
> > -	if (INTEL_GEN(dev_priv) >= 5)
> > +	if (INTEL_GEN(i915) >= 5)
> >   		I915_WRITE(ILK_DPFC_CB_BASE, fbc->compressed_fb.start);
> > -	else if (IS_GM45(dev_priv)) {
> > +	else if (IS_GM45(i915)) {
> >   		I915_WRITE(DPFC_CB_BASE, fbc->compressed_fb.start);
> >   	} else {
> >   		compressed_llb = kzalloc(sizeof(*compressed_llb), GFP_KERNEL);
> >   		if (!compressed_llb)
> >   			goto err_fb;
> > -		ret = i915_gem_stolen_insert_node(dev_priv, compressed_llb,
> > +		ret = i915_gem_stolen_insert_node(i915, compressed_llb,
> >   						  4096, 4096);
> >   		if (ret)
> >   			goto err_fb;
> >   		fbc->compressed_llb = compressed_llb;
> > -		GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start,
> > +		GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start,
> >   					     fbc->compressed_fb.start,
> >   					     U32_MAX));
> > -		GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start,
> > +		GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start,
> >   					     fbc->compressed_llb->start,
> >   					     U32_MAX));
> >   		I915_WRITE(FBC_CFB_BASE,
> > -			   dev_priv->dsm.start + fbc->compressed_fb.start);
> > +			   i915->dsm.start + fbc->compressed_fb.start);
> >   		I915_WRITE(FBC_LL_BASE,
> > -			   dev_priv->dsm.start + compressed_llb->start);
> > +			   i915->dsm.start + compressed_llb->start);
> >   	}
> >   	DRM_DEBUG_KMS("reserved %llu bytes of contiguous stolen space for FBC, threshold: %d\n",
> > @@ -542,39 +542,39 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc)
> >   err_fb:
> >   	kfree(compressed_llb);
> > -	i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb);
> > +	i915_gem_stolen_remove_node(i915, &fbc->compressed_fb);
> >   err_llb:
> > -	if (drm_mm_initialized(&dev_priv->mm.stolen))
> > +	if (drm_mm_initialized(&i915->mm.stolen))
> >   		pr_info_once("drm: not enough stolen space for compressed buffer (need %d more bytes), disabling. Hint: you may be able to increase stolen memory size in the BIOS to avoid this.\n", size);
> >   	return -ENOSPC;
> >   }
> > -static void __intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv)
> > +static void __intel_fbc_cleanup_cfb(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	if (drm_mm_node_allocated(&fbc->compressed_fb))
> > -		i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb);
> > +		i915_gem_stolen_remove_node(i915, &fbc->compressed_fb);
> >   	if (fbc->compressed_llb) {
> > -		i915_gem_stolen_remove_node(dev_priv, fbc->compressed_llb);
> > +		i915_gem_stolen_remove_node(i915, fbc->compressed_llb);
> >   		kfree(fbc->compressed_llb);
> >   	}
> >   }
> > -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv)
> > +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> > -	__intel_fbc_cleanup_cfb(dev_priv);
> > +	__intel_fbc_cleanup_cfb(i915);
> >   	mutex_unlock(&fbc->lock);
> >   }
> > -static bool stride_is_valid(struct drm_i915_private *dev_priv,
> > +static bool stride_is_valid(struct drm_i915_private *i915,
> >   			    unsigned int stride)
> >   {
> >   	/* This should have been caught earlier. */
> > @@ -585,10 +585,10 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv,
> >   	if (stride < 512)
> >   		return false;
> > -	if (IS_GEN(dev_priv, 2) || IS_GEN(dev_priv, 3))
> > +	if (IS_GEN(i915, 2) || IS_GEN(i915, 3))
> >   		return stride == 4096 || stride == 8192;
> > -	if (IS_GEN(dev_priv, 4) && !IS_G4X(dev_priv) && stride < 2048)
> > +	if (IS_GEN(i915, 4) && !IS_G4X(i915) && stride < 2048)
> >   		return false;
> >   	if (stride > 16384)
> > @@ -597,7 +597,7 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> > +static bool pixel_format_is_valid(struct drm_i915_private *i915,
> >   				  u32 pixel_format)
> >   {
> >   	switch (pixel_format) {
> > @@ -607,10 +607,10 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> >   	case DRM_FORMAT_XRGB1555:
> >   	case DRM_FORMAT_RGB565:
> >   		/* 16bpp not supported on gen2 */
> > -		if (IS_GEN(dev_priv, 2))
> > +		if (IS_GEN(i915, 2))
> >   			return false;
> >   		/* WaFbcOnly1to1Ratio:ctg */
> > -		if (IS_G4X(dev_priv))
> > +		if (IS_G4X(i915))
> >   			return false;
> >   		return true;
> >   	default:
> > @@ -626,17 +626,17 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv,
> >    */
> >   static bool intel_fbc_hw_tracking_covers_screen(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	unsigned int effective_w, effective_h, max_w, max_h;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> >   		max_w = 5120;
> >   		max_h = 4096;
> > -	} else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv)) {
> > +	} else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915)) {
> >   		max_w = 4096;
> >   		max_h = 4096;
> > -	} else if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) {
> > +	} else if (IS_G4X(i915) || INTEL_GEN(i915) >= 5) {
> >   		max_w = 4096;
> >   		max_h = 2048;
> >   	} else {
> > @@ -656,8 +656,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
> >   					 struct intel_crtc_state *crtc_state,
> >   					 struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct intel_fbc_state_cache *cache = &fbc->state_cache;
> >   	struct drm_framebuffer *fb = plane_state->base.fb;
> > @@ -665,7 +665,7 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
> >   	cache->flags = 0;
> >   	cache->crtc.mode_flags = crtc_state->base.adjusted_mode.flags;
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		cache->crtc.hsw_bdw_pixel_rate = crtc_state->pixel_rate;
> >   	cache->plane.rotation = plane_state->base.rotation;
> > @@ -697,8 +697,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc,
> >   static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct intel_fbc_state_cache *cache = &fbc->state_cache;
> >   	/* We don't need to use a state cache here since this information is
> > @@ -741,18 +741,18 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   		fbc->no_fbc_reason = "framebuffer not tiled or fenced";
> >   		return false;
> >   	}
> > -	if (INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv) &&
> > +	if (INTEL_GEN(i915) <= 4 && !IS_G4X(i915) &&
> >   	    cache->plane.rotation != DRM_MODE_ROTATE_0) {
> >   		fbc->no_fbc_reason = "rotation unsupported";
> >   		return false;
> >   	}
> > -	if (!stride_is_valid(dev_priv, cache->fb.stride)) {
> > +	if (!stride_is_valid(i915, cache->fb.stride)) {
> >   		fbc->no_fbc_reason = "framebuffer stride not supported";
> >   		return false;
> >   	}
> > -	if (!pixel_format_is_valid(dev_priv, cache->fb.format->format)) {
> > +	if (!pixel_format_is_valid(i915, cache->fb.format->format)) {
> >   		fbc->no_fbc_reason = "pixel format is invalid";
> >   		return false;
> >   	}
> > @@ -764,8 +764,8 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   	}
> >   	/* WaFbcExceedCdClockThreshold:hsw,bdw */
> > -	if ((IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) &&
> > -	    cache->crtc.hsw_bdw_pixel_rate >= dev_priv->cdclk.hw.cdclk * 95 / 100) {
> > +	if ((IS_HASWELL(i915) || IS_BROADWELL(i915)) &&
> > +	    cache->crtc.hsw_bdw_pixel_rate >= i915->cdclk.hw.cdclk * 95 / 100) {
> >   		fbc->no_fbc_reason = "pixel rate is too big";
> >   		return false;
> >   	}
> > @@ -780,7 +780,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   	 * we didn't get any invalidate/deactivate calls, but this would require
> >   	 * a lot of tracking just for a specific case. If we conclude it's an
> >   	 * important case, we can implement it later. */
> > -	if (intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache) >
> > +	if (intel_fbc_calculate_cfb_size(i915, &fbc->state_cache) >
> >   	    fbc->compressed_fb.size * fbc->threshold) {
> >   		fbc->no_fbc_reason = "CFB requirements changed";
> >   		return false;
> > @@ -791,7 +791,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   	 * having a Y offset that isn't divisible by 4 causes FIFO underrun
> >   	 * and screen flicker.
> >   	 */
> > -	if (IS_GEN_RANGE(dev_priv, 9, 10) &&
> > +	if (IS_GEN_RANGE(i915, 9, 10) &&
> >   	    (fbc->state_cache.plane.adjusted_y & 3)) {
> >   		fbc->no_fbc_reason = "plane Y offset is misaligned";
> >   		return false;
> > @@ -800,11 +800,11 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc)
> >   	return true;
> >   }
> > -static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv)
> > +static bool intel_fbc_can_enable(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (intel_vgpu_active(dev_priv)) {
> > +	if (intel_vgpu_active(i915)) {
> >   		fbc->no_fbc_reason = "VGPU is active";
> >   		return false;
> >   	}
> > @@ -825,8 +825,8 @@ static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv)
> >   static void intel_fbc_get_reg_params(struct intel_crtc *crtc,
> >   				     struct intel_fbc_reg_params *params)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct intel_fbc_state_cache *cache = &fbc->state_cache;
> >   	/* Since all our fields are integer types, use memset here so the
> > @@ -844,9 +844,9 @@ static void intel_fbc_get_reg_params(struct intel_crtc *crtc,
> >   	params->fb.format = cache->fb.format;
> >   	params->fb.stride = cache->fb.stride;
> > -	params->cfb_size = intel_fbc_calculate_cfb_size(dev_priv, cache);
> > +	params->cfb_size = intel_fbc_calculate_cfb_size(i915, cache);
> > -	if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915))
> >   		params->gen9_wa_cfb_stride = DIV_ROUND_UP(cache->plane.src_w,
> >   						32 * fbc->threshold) * 8;
> >   }
> > @@ -855,11 +855,11 @@ void intel_fbc_pre_update(struct intel_crtc *crtc,
> >   			  struct intel_crtc_state *crtc_state,
> >   			  struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	const char *reason = "update pending";
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> > @@ -876,21 +876,21 @@ void intel_fbc_pre_update(struct intel_crtc *crtc,
> >   	fbc->flip_pending = true;
> >   deactivate:
> > -	intel_fbc_deactivate(dev_priv, reason);
> > +	intel_fbc_deactivate(i915, reason);
> >   unlock:
> >   	mutex_unlock(&fbc->lock);
> >   }
> >   /**
> >    * __intel_fbc_disable - disable FBC
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This is the low level function that actually disables FBC. Callers should
> >    * grab the FBC lock.
> >    */
> > -static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
> > +static void __intel_fbc_disable(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct intel_crtc *crtc = fbc->crtc;
> >   	WARN_ON(!mutex_is_locked(&fbc->lock));
> > @@ -899,7 +899,7 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_KMS("Disabling FBC on pipe %c\n", pipe_name(crtc->pipe));
> > -	__intel_fbc_cleanup_cfb(dev_priv);
> > +	__intel_fbc_cleanup_cfb(i915);
> >   	fbc->enabled = false;
> >   	fbc->crtc = NULL;
> > @@ -907,8 +907,8 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv)
> >   static void __intel_fbc_post_update(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	WARN_ON(!mutex_is_locked(&fbc->lock));
> > @@ -919,8 +919,8 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc)
> >   	WARN_ON(fbc->active);
> >   	if (!i915_modparams.enable_fbc) {
> > -		intel_fbc_deactivate(dev_priv, "disabled at runtime per module param");
> > -		__intel_fbc_disable(dev_priv);
> > +		intel_fbc_deactivate(i915, "disabled at runtime per module param");
> > +		__intel_fbc_disable(i915);
> >   		return;
> >   	}
> > @@ -931,18 +931,18 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc)
> >   		return;
> >   	if (!fbc->busy_bits) {
> > -		intel_fbc_deactivate(dev_priv, "FBC enabled (active or scheduled)");
> > -		intel_fbc_hw_activate(dev_priv);
> > +		intel_fbc_deactivate(i915, "FBC enabled (active or scheduled)");
> > +		intel_fbc_hw_activate(i915);
> >   	} else
> > -		intel_fbc_deactivate(dev_priv, "frontbuffer write");
> > +		intel_fbc_deactivate(i915, "frontbuffer write");
> >   }
> >   void intel_fbc_post_update(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> > @@ -958,13 +958,13 @@ static unsigned int intel_fbc_get_frontbuffer_bit(struct intel_fbc *fbc)
> >   		return fbc->possible_framebuffer_bits;
> >   }
> > -void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> > +void intel_fbc_invalidate(struct drm_i915_private *i915,
> >   			  unsigned int frontbuffer_bits,
> >   			  enum fb_op_origin origin)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	if (origin == ORIGIN_GTT || origin == ORIGIN_FLIP)
> > @@ -975,17 +975,17 @@ void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> >   	fbc->busy_bits |= intel_fbc_get_frontbuffer_bit(fbc) & frontbuffer_bits;
> >   	if (fbc->enabled && fbc->busy_bits)
> > -		intel_fbc_deactivate(dev_priv, "frontbuffer write");
> > +		intel_fbc_deactivate(i915, "frontbuffer write");
> >   	mutex_unlock(&fbc->lock);
> >   }
> > -void intel_fbc_flush(struct drm_i915_private *dev_priv,
> > +void intel_fbc_flush(struct drm_i915_private *i915,
> >   		     unsigned int frontbuffer_bits, enum fb_op_origin origin)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> > @@ -998,7 +998,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
> >   	if (!fbc->busy_bits && fbc->enabled &&
> >   	    (frontbuffer_bits & intel_fbc_get_frontbuffer_bit(fbc))) {
> >   		if (fbc->active)
> > -			intel_fbc_recompress(dev_priv);
> > +			intel_fbc_recompress(i915);
> >   		else if (!fbc->flip_pending)
> >   			__intel_fbc_post_update(fbc->crtc);
> >   	}
> > @@ -1009,7 +1009,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_fbc_choose_crtc - select a CRTC to enable FBC on
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @state: the atomic state structure
> >    *
> >    * This function looks at the proposed state for CRTCs and planes, then chooses
> > @@ -1017,12 +1017,12 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv,
> >    * true.
> >    *
> >    * Later, intel_fbc_enable is going to look for state->enable_fbc and then maybe
> > - * enable FBC for the chosen CRTC. If it does, it will set dev_priv->fbc.crtc.
> > + * enable FBC for the chosen CRTC. If it does, it will set i915->fbc.crtc.
> >    */
> > -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> > +void intel_fbc_choose_crtc(struct drm_i915_private *i915,
> >   			   struct intel_atomic_state *state)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	struct intel_plane *plane;
> >   	struct intel_plane_state *plane_state;
> >   	bool crtc_chosen = false;
> > @@ -1035,7 +1035,7 @@ void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> >   	    !intel_atomic_get_new_crtc_state(state, fbc->crtc))
> >   		goto out;
> > -	if (!intel_fbc_can_enable(dev_priv))
> > +	if (!intel_fbc_can_enable(i915))
> >   		goto out;
> >   	/* Simply choose the first CRTC that is compatible and has a visible
> > @@ -1081,10 +1081,10 @@ void intel_fbc_enable(struct intel_crtc *crtc,
> >   		      struct intel_crtc_state *crtc_state,
> >   		      struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> > @@ -1127,44 +1127,44 @@ void intel_fbc_enable(struct intel_crtc *crtc,
> >    */
> >   void intel_fbc_disable(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> >   	if (fbc->crtc == crtc)
> > -		__intel_fbc_disable(dev_priv);
> > +		__intel_fbc_disable(i915);
> >   	mutex_unlock(&fbc->lock);
> >   }
> >   /**
> >    * intel_fbc_global_disable - globally disable FBC
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function disables FBC regardless of which CRTC is associated with it.
> >    */
> > -void intel_fbc_global_disable(struct drm_i915_private *dev_priv)
> > +void intel_fbc_global_disable(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	mutex_lock(&fbc->lock);
> >   	if (fbc->enabled) {
> >   		WARN_ON(fbc->crtc->active);
> > -		__intel_fbc_disable(dev_priv);
> > +		__intel_fbc_disable(i915);
> >   	}
> >   	mutex_unlock(&fbc->lock);
> >   }
> >   static void intel_fbc_underrun_work_fn(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private, fbc.underrun_work);
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	mutex_lock(&fbc->lock);
> > @@ -1175,42 +1175,42 @@ static void intel_fbc_underrun_work_fn(struct work_struct *work)
> >   	DRM_DEBUG_KMS("Disabling FBC due to FIFO underrun.\n");
> >   	fbc->underrun_detected = true;
> > -	intel_fbc_deactivate(dev_priv, "FIFO underrun");
> > +	intel_fbc_deactivate(i915, "FIFO underrun");
> >   out:
> >   	mutex_unlock(&fbc->lock);
> >   }
> >   /*
> >    * intel_fbc_reset_underrun - reset FBC fifo underrun status.
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * See intel_fbc_handle_fifo_underrun_irq(). For automated testing we
> >    * want to re-enable FBC after an underrun to increase test coverage.
> >    */
> > -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv)
> > +int intel_fbc_reset_underrun(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	cancel_work_sync(&dev_priv->fbc.underrun_work);
> > +	cancel_work_sync(&i915->fbc.underrun_work);
> > -	ret = mutex_lock_interruptible(&dev_priv->fbc.lock);
> > +	ret = mutex_lock_interruptible(&i915->fbc.lock);
> >   	if (ret)
> >   		return ret;
> > -	if (dev_priv->fbc.underrun_detected) {
> > +	if (i915->fbc.underrun_detected) {
> >   		DRM_DEBUG_KMS("Re-allowing FBC after fifo underrun\n");
> > -		dev_priv->fbc.no_fbc_reason = "FIFO underrun cleared";
> > +		i915->fbc.no_fbc_reason = "FIFO underrun cleared";
> >   	}
> > -	dev_priv->fbc.underrun_detected = false;
> > -	mutex_unlock(&dev_priv->fbc.lock);
> > +	i915->fbc.underrun_detected = false;
> > +	mutex_unlock(&i915->fbc.lock);
> >   	return 0;
> >   }
> >   /**
> >    * intel_fbc_handle_fifo_underrun_irq - disable FBC when we get a FIFO underrun
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Without FBC, most underruns are harmless and don't really cause too many
> >    * problems, except for an annoying message on dmesg. With FBC, underruns can
> > @@ -1222,11 +1222,11 @@ int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv)
> >    *
> >    * This function is called from the IRQ handler.
> >    */
> > -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv)
> > +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> > -	if (!fbc_supported(dev_priv))
> > +	if (!fbc_supported(i915))
> >   		return;
> >   	/* There's no guarantee that underrun_detected won't be set to true
> > @@ -1243,24 +1243,24 @@ void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_fbc_init_pipe_state - initialize FBC's CRTC visibility tracking
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * The FBC code needs to track CRTC visibility since the older platforms can't
> >    * have FBC enabled while multiple pipes are used. This function does the
> >    * initial setup at driver load to make sure FBC is matching the real hardware.
> >    */
> > -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv)
> > +void intel_fbc_init_pipe_state(struct drm_i915_private *i915)
> >   {
> >   	struct intel_crtc *crtc;
> >   	/* Don't even bother tracking anything if we don't need. */
> > -	if (!no_fbc_on_multiple_pipes(dev_priv))
> > +	if (!no_fbc_on_multiple_pipes(i915))
> >   		return;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc)
> > +	for_each_intel_crtc(&i915->drm, crtc)
> >   		if (intel_crtc_active(crtc) &&
> >   		    crtc->base.primary->state->visible)
> > -			dev_priv->fbc.visible_pipes_mask |= (1 << crtc->pipe);
> > +			i915->fbc.visible_pipes_mask |= (1 << crtc->pipe);
> >   }
> >   /*
> > @@ -1272,29 +1272,29 @@ void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv)
> >    * space to change the value during runtime without sanitizing it again. IGT
> >    * relies on being able to change i915.enable_fbc at runtime.
> >    */
> > -static int intel_sanitize_fbc_option(struct drm_i915_private *dev_priv)
> > +static int intel_sanitize_fbc_option(struct drm_i915_private *i915)
> >   {
> >   	if (i915_modparams.enable_fbc >= 0)
> >   		return !!i915_modparams.enable_fbc;
> > -	if (!HAS_FBC(dev_priv))
> > +	if (!HAS_FBC(i915))
> >   		return 0;
> >   	/* https://bugs.freedesktop.org/show_bug.cgi?id=108085 */
> > -	if (IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEMINILAKE(i915))
> >   		return 0;
> > -	if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9)
> > +	if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9)
> >   		return 1;
> >   	return 0;
> >   }
> > -static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv)
> > +static bool need_fbc_vtd_wa(struct drm_i915_private *i915)
> >   {
> >   	/* WaFbcTurnOffFbcWhenHyperVisorIsUsed:skl,bxt */
> >   	if (intel_vtd_active() &&
> > -	    (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv))) {
> > +	    (IS_SKYLAKE(i915) || IS_BROXTON(i915))) {
> >   		DRM_INFO("Disabling framebuffer compression (FBC) to prevent screen flicker with VT-d enabled\n");
> >   		return true;
> >   	}
> > @@ -1304,38 +1304,38 @@ static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_fbc_init - Initialize FBC
> > - * @dev_priv: the i915 device
> > + * @i915: the i915 device
> >    *
> >    * This function might be called during PM init process.
> >    */
> > -void intel_fbc_init(struct drm_i915_private *dev_priv)
> > +void intel_fbc_init(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbc *fbc = &dev_priv->fbc;
> > +	struct intel_fbc *fbc = &i915->fbc;
> >   	INIT_WORK(&fbc->underrun_work, intel_fbc_underrun_work_fn);
> >   	mutex_init(&fbc->lock);
> >   	fbc->enabled = false;
> >   	fbc->active = false;
> > -	if (need_fbc_vtd_wa(dev_priv))
> > -		mkwrite_device_info(dev_priv)->display.has_fbc = false;
> > +	if (need_fbc_vtd_wa(i915))
> > +		mkwrite_device_info(i915)->display.has_fbc = false;
> > -	i915_modparams.enable_fbc = intel_sanitize_fbc_option(dev_priv);
> > +	i915_modparams.enable_fbc = intel_sanitize_fbc_option(i915);
> >   	DRM_DEBUG_KMS("Sanitized enable_fbc value: %d\n",
> >   		      i915_modparams.enable_fbc);
> > -	if (!HAS_FBC(dev_priv)) {
> > +	if (!HAS_FBC(i915)) {
> >   		fbc->no_fbc_reason = "unsupported by this chipset";
> >   		return;
> >   	}
> >   	/* This value was pulled out of someone's hat */
> > -	if (INTEL_GEN(dev_priv) <= 4 && !IS_GM45(dev_priv))
> > +	if (INTEL_GEN(i915) <= 4 && !IS_GM45(i915))
> >   		I915_WRITE(FBC_CONTROL, 500 << FBC_CTL_INTERVAL_SHIFT);
> >   	/* We still don't have any sort of hardware state readout for FBC, so
> >   	 * deactivate it in case the BIOS activated it to make sure software
> >   	 * matches the hardware state. */
> > -	if (intel_fbc_hw_is_active(dev_priv))
> > -		intel_fbc_hw_deactivate(dev_priv);
> > +	if (intel_fbc_hw_is_active(i915))
> > +		intel_fbc_hw_deactivate(i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_fbc.h b/drivers/gpu/drm/i915/intel_fbc.h
> > index 50272eda8d43..b7dc6cb43497 100644
> > --- a/drivers/gpu/drm/i915/intel_fbc.h
> > +++ b/drivers/gpu/drm/i915/intel_fbc.h
> > @@ -16,27 +16,27 @@ struct intel_crtc;
> >   struct intel_crtc_state;
> >   struct intel_plane_state;
> > -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv,
> > +void intel_fbc_choose_crtc(struct drm_i915_private *i915,
> >   			   struct intel_atomic_state *state);
> > -bool intel_fbc_is_active(struct drm_i915_private *dev_priv);
> > +bool intel_fbc_is_active(struct drm_i915_private *i915);
> >   void intel_fbc_pre_update(struct intel_crtc *crtc,
> >   			  struct intel_crtc_state *crtc_state,
> >   			  struct intel_plane_state *plane_state);
> >   void intel_fbc_post_update(struct intel_crtc *crtc);
> > -void intel_fbc_init(struct drm_i915_private *dev_priv);
> > -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv);
> > +void intel_fbc_init(struct drm_i915_private *i915);
> > +void intel_fbc_init_pipe_state(struct drm_i915_private *i915);
> >   void intel_fbc_enable(struct intel_crtc *crtc,
> >   		      struct intel_crtc_state *crtc_state,
> >   		      struct intel_plane_state *plane_state);
> >   void intel_fbc_disable(struct intel_crtc *crtc);
> > -void intel_fbc_global_disable(struct drm_i915_private *dev_priv);
> > -void intel_fbc_invalidate(struct drm_i915_private *dev_priv,
> > +void intel_fbc_global_disable(struct drm_i915_private *i915);
> > +void intel_fbc_invalidate(struct drm_i915_private *i915,
> >   			  unsigned int frontbuffer_bits,
> >   			  enum fb_op_origin origin);
> > -void intel_fbc_flush(struct drm_i915_private *dev_priv,
> > +void intel_fbc_flush(struct drm_i915_private *i915,
> >   		     unsigned int frontbuffer_bits, enum fb_op_origin origin);
> > -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv);
> > -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv);
> > -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv);
> > +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915);
> > +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915);
> > +int intel_fbc_reset_underrun(struct drm_i915_private *i915);
> >   #endif /* __INTEL_FBC_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_fbdev.c b/drivers/gpu/drm/i915/intel_fbdev.c
> > index 0d3a6fa674e6..93dee9ff5a58 100644
> > --- a/drivers/gpu/drm/i915/intel_fbdev.c
> > +++ b/drivers/gpu/drm/i915/intel_fbdev.c
> > @@ -117,7 +117,7 @@ static int intelfb_alloc(struct drm_fb_helper *helper,
> >   		container_of(helper, struct intel_fbdev, helper);
> >   	struct drm_framebuffer *fb;
> >   	struct drm_device *dev = helper->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_mode_fb_cmd2 mode_cmd = {};
> >   	struct drm_i915_gem_object *obj;
> >   	int size, ret;
> > @@ -141,10 +141,10 @@ static int intelfb_alloc(struct drm_fb_helper *helper,
> >   	 * important and we should probably use that space with FBC or other
> >   	 * features. */
> >   	obj = NULL;
> > -	if (size * 2 < dev_priv->stolen_usable_size)
> > -		obj = i915_gem_object_create_stolen(dev_priv, size);
> > +	if (size * 2 < i915->stolen_usable_size)
> > +		obj = i915_gem_object_create_stolen(i915, size);
> >   	if (obj == NULL)
> > -		obj = i915_gem_object_create_shmem(dev_priv, size);
> > +		obj = i915_gem_object_create_shmem(i915, size);
> >   	if (IS_ERR(obj)) {
> >   		DRM_ERROR("failed to allocate framebuffer\n");
> >   		ret = PTR_ERR(obj);
> > @@ -174,9 +174,9 @@ static int intelfb_create(struct drm_fb_helper *helper,
> >   		container_of(helper, struct intel_fbdev, helper);
> >   	struct intel_framebuffer *intel_fb = ifbdev->fb;
> >   	struct drm_device *dev = helper->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > -	struct i915_ggtt *ggtt = &dev_priv->ggtt;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct pci_dev *pdev = i915->drm.pdev;
> > +	struct i915_ggtt *ggtt = &i915->ggtt;
> >   	const struct i915_ggtt_view view = {
> >   		.type = I915_GGTT_VIEW_NORMAL,
> >   	};
> > @@ -213,7 +213,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> >   	}
> >   	mutex_lock(&dev->struct_mutex);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> >   	/* Pin the GGTT vma for our access via info->screen_base.
> >   	 * This also validates that any existing fb inherited from the
> > @@ -272,7 +272,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> >   	ifbdev->vma = vma;
> >   	ifbdev->vma_flags = flags;
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	mutex_unlock(&dev->struct_mutex);
> >   	vga_switcheroo_client_fb_set(pdev, info);
> >   	return 0;
> > @@ -280,7 +280,7 @@ static int intelfb_create(struct drm_fb_helper *helper,
> >   out_unpin:
> >   	intel_unpin_fb_vma(vma, flags);
> >   out_unlock:
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   	mutex_unlock(&dev->struct_mutex);
> >   	return ret;
> >   }
> > @@ -447,11 +447,11 @@ static void intel_fbdev_suspend_worker(struct work_struct *work)
> >   int intel_fbdev_init(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_fbdev *ifbdev;
> >   	int ret;
> > -	if (WARN_ON(!HAS_DISPLAY(dev_priv)))
> > +	if (WARN_ON(!HAS_DISPLAY(i915)))
> >   		return -ENODEV;
> >   	ifbdev = kzalloc(sizeof(struct intel_fbdev), GFP_KERNEL);
> > @@ -470,8 +470,8 @@ int intel_fbdev_init(struct drm_device *dev)
> >   		return ret;
> >   	}
> > -	dev_priv->fbdev = ifbdev;
> > -	INIT_WORK(&dev_priv->fbdev_suspend_work, intel_fbdev_suspend_worker);
> > +	i915->fbdev = ifbdev;
> > +	INIT_WORK(&i915->fbdev_suspend_work, intel_fbdev_suspend_worker);
> >   	drm_fb_helper_single_add_all_connectors(&ifbdev->helper);
> > @@ -508,23 +508,23 @@ static void intel_fbdev_sync(struct intel_fbdev *ifbdev)
> >   	ifbdev->cookie = 0;
> >   }
> > -void intel_fbdev_unregister(struct drm_i915_private *dev_priv)
> > +void intel_fbdev_unregister(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbdev *ifbdev = dev_priv->fbdev;
> > +	struct intel_fbdev *ifbdev = i915->fbdev;
> >   	if (!ifbdev)
> >   		return;
> > -	cancel_work_sync(&dev_priv->fbdev_suspend_work);
> > +	cancel_work_sync(&i915->fbdev_suspend_work);
> >   	if (!current_is_async())
> >   		intel_fbdev_sync(ifbdev);
> >   	drm_fb_helper_unregister_fbi(&ifbdev->helper);
> >   }
> > -void intel_fbdev_fini(struct drm_i915_private *dev_priv)
> > +void intel_fbdev_fini(struct drm_i915_private *i915)
> >   {
> > -	struct intel_fbdev *ifbdev = fetch_and_zero(&dev_priv->fbdev);
> > +	struct intel_fbdev *ifbdev = fetch_and_zero(&i915->fbdev);
> >   	if (!ifbdev)
> >   		return;
> > @@ -554,8 +554,8 @@ static void intel_fbdev_hpd_set_suspend(struct intel_fbdev *ifbdev, int state)
> >   void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_fbdev *ifbdev = dev_priv->fbdev;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_fbdev *ifbdev = i915->fbdev;
> >   	struct fb_info *info;
> >   	if (!ifbdev || !ifbdev->vma)
> > @@ -572,7 +572,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous
> >   		 * ourselves, so only flush outstanding work upon suspend!
> >   		 */
> >   		if (state != FBINFO_STATE_RUNNING)
> > -			flush_work(&dev_priv->fbdev_suspend_work);
> > +			flush_work(&i915->fbdev_suspend_work);
> >   		console_lock();
> >   	} else {
> > @@ -586,7 +586,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous
> >   			/* Don't block our own workqueue as this can
> >   			 * be run in parallel with other i915.ko tasks.
> >   			 */
> > -			schedule_work(&dev_priv->fbdev_suspend_work);
> > +			schedule_work(&i915->fbdev_suspend_work);
> >   			return;
> >   		}
> >   	}
> > diff --git a/drivers/gpu/drm/i915/intel_fbdev.h b/drivers/gpu/drm/i915/intel_fbdev.h
> > index de7c84250eb5..4d0c8c851a5e 100644
> > --- a/drivers/gpu/drm/i915/intel_fbdev.h
> > +++ b/drivers/gpu/drm/i915/intel_fbdev.h
> > @@ -14,8 +14,8 @@ struct drm_i915_private;
> >   #ifdef CONFIG_DRM_FBDEV_EMULATION
> >   int intel_fbdev_init(struct drm_device *dev);
> >   void intel_fbdev_initial_config_async(struct drm_device *dev);
> > -void intel_fbdev_unregister(struct drm_i915_private *dev_priv);
> > -void intel_fbdev_fini(struct drm_i915_private *dev_priv);
> > +void intel_fbdev_unregister(struct drm_i915_private *i915);
> > +void intel_fbdev_fini(struct drm_i915_private *i915);
> >   void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous);
> >   void intel_fbdev_output_poll_changed(struct drm_device *dev);
> >   void intel_fbdev_restore_mode(struct drm_device *dev);
> > @@ -29,11 +29,11 @@ static inline void intel_fbdev_initial_config_async(struct drm_device *dev)
> >   {
> >   }
> > -static inline void intel_fbdev_unregister(struct drm_i915_private *dev_priv)
> > +static inline void intel_fbdev_unregister(struct drm_i915_private *i915)
> >   {
> >   }
> > -static inline void intel_fbdev_fini(struct drm_i915_private *dev_priv)
> > +static inline void intel_fbdev_fini(struct drm_i915_private *i915)
> >   {
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.c b/drivers/gpu/drm/i915/intel_fifo_underrun.c
> > index 8545ad32bb50..08f092addae2 100644
> > --- a/drivers/gpu/drm/i915/intel_fifo_underrun.c
> > +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.c
> > @@ -52,14 +52,14 @@
> >   static bool ivb_can_enable_err_int(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *crtc;
> >   	enum pipe pipe;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	for_each_pipe(dev_priv, pipe) {
> > -		crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	for_each_pipe(i915, pipe) {
> > +		crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   		if (crtc->cpu_fifo_underrun_disabled)
> >   			return false;
> > @@ -70,14 +70,14 @@ static bool ivb_can_enable_err_int(struct drm_device *dev)
> >   static bool cpt_can_enable_serr_int(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum pipe pipe;
> >   	struct intel_crtc *crtc;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> > -	for_each_pipe(dev_priv, pipe) {
> > -		crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	for_each_pipe(i915, pipe) {
> > +		crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   		if (crtc->pch_fifo_underrun_disabled)
> >   			return false;
> > @@ -88,20 +88,20 @@ static bool cpt_can_enable_serr_int(struct drm_device *dev)
> >   static void i9xx_check_fifo_underruns(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	i915_reg_t reg = PIPESTAT(crtc->pipe);
> >   	u32 enable_mask;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	if ((I915_READ(reg) & PIPE_FIFO_UNDERRUN_STATUS) == 0)
> >   		return;
> > -	enable_mask = i915_pipestat_enable_mask(dev_priv, crtc->pipe);
> > +	enable_mask = i915_pipestat_enable_mask(i915, crtc->pipe);
> >   	I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS);
> >   	POSTING_READ(reg);
> > -	trace_intel_cpu_fifo_underrun(dev_priv, crtc->pipe);
> > +	trace_intel_cpu_fifo_underrun(i915, crtc->pipe);
> >   	DRM_ERROR("pipe %c underrun\n", pipe_name(crtc->pipe));
> >   }
> > @@ -109,13 +109,13 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev,
> >   					     enum pipe pipe,
> >   					     bool enable, bool old)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	i915_reg_t reg = PIPESTAT(pipe);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	if (enable) {
> > -		u32 enable_mask = i915_pipestat_enable_mask(dev_priv, pipe);
> > +		u32 enable_mask = i915_pipestat_enable_mask(i915, pipe);
> >   		I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS);
> >   		POSTING_READ(reg);
> > @@ -128,23 +128,23 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev,
> >   static void ironlake_set_fifo_underrun_reporting(struct drm_device *dev,
> >   						 enum pipe pipe, bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 bit = (pipe == PIPE_A) ?
> >   		DE_PIPEA_FIFO_UNDERRUN : DE_PIPEB_FIFO_UNDERRUN;
> >   	if (enable)
> > -		ilk_enable_display_irq(dev_priv, bit);
> > +		ilk_enable_display_irq(i915, bit);
> >   	else
> > -		ilk_disable_display_irq(dev_priv, bit);
> > +		ilk_disable_display_irq(i915, bit);
> >   }
> >   static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	u32 err_int = I915_READ(GEN7_ERR_INT);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	if ((err_int & ERR_INT_FIFO_UNDERRUN(pipe)) == 0)
> >   		return;
> > @@ -152,7 +152,7 @@ static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc)
> >   	I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe));
> >   	POSTING_READ(GEN7_ERR_INT);
> > -	trace_intel_cpu_fifo_underrun(dev_priv, pipe);
> > +	trace_intel_cpu_fifo_underrun(i915, pipe);
> >   	DRM_ERROR("fifo underrun on pipe %c\n", pipe_name(pipe));
> >   }
> > @@ -160,16 +160,16 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev,
> >   						  enum pipe pipe,
> >   						  bool enable, bool old)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	if (enable) {
> >   		I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe));
> >   		if (!ivb_can_enable_err_int(dev))
> >   			return;
> > -		ilk_enable_display_irq(dev_priv, DE_ERR_INT_IVB);
> > +		ilk_enable_display_irq(i915, DE_ERR_INT_IVB);
> >   	} else {
> > -		ilk_disable_display_irq(dev_priv, DE_ERR_INT_IVB);
> > +		ilk_disable_display_irq(i915, DE_ERR_INT_IVB);
> >   		if (old &&
> >   		    I915_READ(GEN7_ERR_INT) & ERR_INT_FIFO_UNDERRUN(pipe)) {
> > @@ -182,35 +182,35 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev,
> >   static void broadwell_set_fifo_underrun_reporting(struct drm_device *dev,
> >   						  enum pipe pipe, bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	if (enable)
> > -		bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> > +		bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> >   	else
> > -		bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> > +		bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN);
> >   }
> >   static void ibx_set_fifo_underrun_reporting(struct drm_device *dev,
> >   					    enum pipe pch_transcoder,
> >   					    bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 bit = (pch_transcoder == PIPE_A) ?
> >   		SDE_TRANSA_FIFO_UNDER : SDE_TRANSB_FIFO_UNDER;
> >   	if (enable)
> > -		ibx_enable_display_interrupt(dev_priv, bit);
> > +		ibx_enable_display_interrupt(i915, bit);
> >   	else
> > -		ibx_disable_display_interrupt(dev_priv, bit);
> > +		ibx_disable_display_interrupt(i915, bit);
> >   }
> >   static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pch_transcoder = crtc->pipe;
> >   	u32 serr_int = I915_READ(SERR_INT);
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	if ((serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) == 0)
> >   		return;
> > @@ -218,7 +218,7 @@ static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
> >   	I915_WRITE(SERR_INT, SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder));
> >   	POSTING_READ(SERR_INT);
> > -	trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
> > +	trace_intel_pch_fifo_underrun(i915, pch_transcoder);
> >   	DRM_ERROR("pch fifo underrun on pch transcoder %c\n",
> >   		  pipe_name(pch_transcoder));
> >   }
> > @@ -227,7 +227,7 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> >   					    enum pipe pch_transcoder,
> >   					    bool enable, bool old)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	if (enable) {
> >   		I915_WRITE(SERR_INT,
> > @@ -236,9 +236,9 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> >   		if (!cpt_can_enable_serr_int(dev))
> >   			return;
> > -		ibx_enable_display_interrupt(dev_priv, SDE_ERROR_CPT);
> > +		ibx_enable_display_interrupt(i915, SDE_ERROR_CPT);
> >   	} else {
> > -		ibx_disable_display_interrupt(dev_priv, SDE_ERROR_CPT);
> > +		ibx_disable_display_interrupt(i915, SDE_ERROR_CPT);
> >   		if (old && I915_READ(SERR_INT) &
> >   		    SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) {
> > @@ -251,22 +251,22 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
> >   static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
> >   						    enum pipe pipe, bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	bool old;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	old = !crtc->cpu_fifo_underrun_disabled;
> >   	crtc->cpu_fifo_underrun_disabled = !enable;
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		i9xx_set_fifo_underrun_reporting(dev, pipe, enable, old);
> > -	else if (IS_GEN_RANGE(dev_priv, 5, 6))
> > +	else if (IS_GEN_RANGE(i915, 5, 6))
> >   		ironlake_set_fifo_underrun_reporting(dev, pipe, enable);
> > -	else if (IS_GEN(dev_priv, 7))
> > +	else if (IS_GEN(i915, 7))
> >   		ivybridge_set_fifo_underrun_reporting(dev, pipe, enable, old);
> > -	else if (INTEL_GEN(dev_priv) >= 8)
> > +	else if (INTEL_GEN(i915) >= 8)
> >   		broadwell_set_fifo_underrun_reporting(dev, pipe, enable);
> >   	return old;
> > @@ -274,7 +274,7 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
> >   /**
> >    * intel_set_cpu_fifo_underrun_reporting - set cpu fifo underrrun reporting state
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @pipe: (CPU) pipe to set state for
> >    * @enable: whether underruns should be reported or not
> >    *
> > @@ -288,23 +288,23 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev,
> >    *
> >    * Returns the previous state of underrun reporting.
> >    */
> > -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> > +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915,
> >   					   enum pipe pipe, bool enable)
> >   {
> >   	unsigned long flags;
> >   	bool ret;
> > -	spin_lock_irqsave(&dev_priv->irq_lock, flags);
> > -	ret = __intel_set_cpu_fifo_underrun_reporting(&dev_priv->drm, pipe,
> > +	spin_lock_irqsave(&i915->irq_lock, flags);
> > +	ret = __intel_set_cpu_fifo_underrun_reporting(&i915->drm, pipe,
> >   						      enable);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
> > +	spin_unlock_irqrestore(&i915->irq_lock, flags);
> >   	return ret;
> >   }
> >   /**
> >    * intel_set_pch_fifo_underrun_reporting - set PCH fifo underrun reporting state
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older)
> >    * @enable: whether underruns should be reported or not
> >    *
> > @@ -316,12 +316,12 @@ bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> >    *
> >    * Returns the previous state of underrun reporting.
> >    */
> > -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> > +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915,
> >   					   enum pipe pch_transcoder,
> >   					   bool enable)
> >   {
> >   	struct intel_crtc *crtc =
> > -		intel_get_crtc_for_pipe(dev_priv, pch_transcoder);
> > +		intel_get_crtc_for_pipe(i915, pch_transcoder);
> >   	unsigned long flags;
> >   	bool old;
> > @@ -334,71 +334,71 @@ bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> >   	 * crtc on LPT won't cause issues.
> >   	 */
> > -	spin_lock_irqsave(&dev_priv->irq_lock, flags);
> > +	spin_lock_irqsave(&i915->irq_lock, flags);
> >   	old = !crtc->pch_fifo_underrun_disabled;
> >   	crtc->pch_fifo_underrun_disabled = !enable;
> > -	if (HAS_PCH_IBX(dev_priv))
> > -		ibx_set_fifo_underrun_reporting(&dev_priv->drm,
> > +	if (HAS_PCH_IBX(i915))
> > +		ibx_set_fifo_underrun_reporting(&i915->drm,
> >   						pch_transcoder,
> >   						enable);
> >   	else
> > -		cpt_set_fifo_underrun_reporting(&dev_priv->drm,
> > +		cpt_set_fifo_underrun_reporting(&i915->drm,
> >   						pch_transcoder,
> >   						enable, old);
> > -	spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
> > +	spin_unlock_irqrestore(&i915->irq_lock, flags);
> >   	return old;
> >   }
> >   /**
> >    * intel_cpu_fifo_underrun_irq_handler - handle CPU fifo underrun interrupt
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @pipe: (CPU) pipe to set state for
> >    *
> >    * This handles a CPU fifo underrun interrupt, generating an underrun warning
> >    * into dmesg if underrun reporting is enabled and then disables the underrun
> >    * interrupt to avoid an irq storm.
> >    */
> > -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> >   					 enum pipe pipe)
> >   {
> > -	struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	/* We may be called too early in init, thanks BIOS! */
> >   	if (crtc == NULL)
> >   		return;
> >   	/* GMCH can't disable fifo underruns, filter them. */
> > -	if (HAS_GMCH(dev_priv) &&
> > +	if (HAS_GMCH(i915) &&
> >   	    crtc->cpu_fifo_underrun_disabled)
> >   		return;
> > -	if (intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false)) {
> > -		trace_intel_cpu_fifo_underrun(dev_priv, pipe);
> > +	if (intel_set_cpu_fifo_underrun_reporting(i915, pipe, false)) {
> > +		trace_intel_cpu_fifo_underrun(i915, pipe);
> >   		DRM_ERROR("CPU pipe %c FIFO underrun\n",
> >   			  pipe_name(pipe));
> >   	}
> > -	intel_fbc_handle_fifo_underrun_irq(dev_priv);
> > +	intel_fbc_handle_fifo_underrun_irq(i915);
> >   }
> >   /**
> >    * intel_pch_fifo_underrun_irq_handler - handle PCH fifo underrun interrupt
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older)
> >    *
> >    * This handles a PCH fifo underrun interrupt, generating an underrun warning
> >    * into dmesg if underrun reporting is enabled and then disables the underrun
> >    * interrupt to avoid an irq storm.
> >    */
> > -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> >   					 enum pipe pch_transcoder)
> >   {
> > -	if (intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder,
> > +	if (intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder,
> >   						  false)) {
> > -		trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
> > +		trace_intel_pch_fifo_underrun(i915, pch_transcoder);
> >   		DRM_ERROR("PCH transcoder %c FIFO underrun\n",
> >   			  pipe_name(pch_transcoder));
> >   	}
> > @@ -406,53 +406,53 @@ void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_check_cpu_fifo_underruns - check for CPU fifo underruns immediately
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Check for CPU fifo underruns immediately. Useful on IVB/HSW where the shared
> >    * error interrupt may have been disabled, and so CPU fifo underruns won't
> >    * necessarily raise an interrupt, and on GMCH platforms where underruns never
> >    * raise an interrupt.
> >    */
> > -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv)
> > +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915)
> >   {
> >   	struct intel_crtc *crtc;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		if (crtc->cpu_fifo_underrun_disabled)
> >   			continue;
> > -		if (HAS_GMCH(dev_priv))
> > +		if (HAS_GMCH(i915))
> >   			i9xx_check_fifo_underruns(crtc);
> > -		else if (IS_GEN(dev_priv, 7))
> > +		else if (IS_GEN(i915, 7))
> >   			ivybridge_check_fifo_underruns(crtc);
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> >   /**
> >    * intel_check_pch_fifo_underruns - check for PCH fifo underruns immediately
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * Check for PCH fifo underruns immediately. Useful on CPT/PPT where the shared
> >    * error interrupt may have been disabled, and so PCH fifo underruns won't
> >    * necessarily raise an interrupt.
> >    */
> > -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv)
> > +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915)
> >   {
> >   	struct intel_crtc *crtc;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		if (crtc->pch_fifo_underrun_disabled)
> >   			continue;
> > -		if (HAS_PCH_CPT(dev_priv))
> > +		if (HAS_PCH_CPT(i915))
> >   			cpt_check_pch_fifo_underruns(crtc);
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.h b/drivers/gpu/drm/i915/intel_fifo_underrun.h
> > index e04f22ac1f49..290909a67de8 100644
> > --- a/drivers/gpu/drm/i915/intel_fifo_underrun.h
> > +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.h
> > @@ -12,16 +12,16 @@
> >   struct drm_i915_private;
> > -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> > +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915,
> >   					   enum pipe pipe, bool enable);
> > -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv,
> > +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915,
> >   					   enum pipe pch_transcoder,
> >   					   bool enable);
> > -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> >   					 enum pipe pipe);
> > -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915,
> >   					 enum pipe pch_transcoder);
> > -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv);
> > -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv);
> > +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915);
> > +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915);
> >   #endif /* __INTEL_FIFO_UNDERRUN_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.c b/drivers/gpu/drm/i915/intel_frontbuffer.c
> > index d6036b9ad16a..161139de3f28 100644
> > --- a/drivers/gpu/drm/i915/intel_frontbuffer.c
> > +++ b/drivers/gpu/drm/i915/intel_frontbuffer.c
> > @@ -66,24 +66,24 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> >   			       enum fb_op_origin origin,
> >   			       unsigned int frontbuffer_bits)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	if (origin == ORIGIN_CS) {
> > -		spin_lock(&dev_priv->fb_tracking.lock);
> > -		dev_priv->fb_tracking.busy_bits |= frontbuffer_bits;
> > -		dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits;
> > -		spin_unlock(&dev_priv->fb_tracking.lock);
> > +		spin_lock(&i915->fb_tracking.lock);
> > +		i915->fb_tracking.busy_bits |= frontbuffer_bits;
> > +		i915->fb_tracking.flip_bits &= ~frontbuffer_bits;
> > +		spin_unlock(&i915->fb_tracking.lock);
> >   	}
> >   	might_sleep();
> > -	intel_psr_invalidate(dev_priv, frontbuffer_bits, origin);
> > -	intel_edp_drrs_invalidate(dev_priv, frontbuffer_bits);
> > -	intel_fbc_invalidate(dev_priv, frontbuffer_bits, origin);
> > +	intel_psr_invalidate(i915, frontbuffer_bits, origin);
> > +	intel_edp_drrs_invalidate(i915, frontbuffer_bits);
> > +	intel_fbc_invalidate(i915, frontbuffer_bits, origin);
> >   }
> >   /**
> >    * intel_frontbuffer_flush - flush frontbuffer
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    * @origin: which operation caused the flush
> >    *
> > @@ -93,45 +93,45 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> >    *
> >    * Can be called without any locks held.
> >    */
> > -static void intel_frontbuffer_flush(struct drm_i915_private *dev_priv,
> > +static void intel_frontbuffer_flush(struct drm_i915_private *i915,
> >   				    unsigned frontbuffer_bits,
> >   				    enum fb_op_origin origin)
> >   {
> >   	/* Delay flushing when rings are still busy.*/
> > -	spin_lock(&dev_priv->fb_tracking.lock);
> > -	frontbuffer_bits &= ~dev_priv->fb_tracking.busy_bits;
> > -	spin_unlock(&dev_priv->fb_tracking.lock);
> > +	spin_lock(&i915->fb_tracking.lock);
> > +	frontbuffer_bits &= ~i915->fb_tracking.busy_bits;
> > +	spin_unlock(&i915->fb_tracking.lock);
> >   	if (!frontbuffer_bits)
> >   		return;
> >   	might_sleep();
> > -	intel_edp_drrs_flush(dev_priv, frontbuffer_bits);
> > -	intel_psr_flush(dev_priv, frontbuffer_bits, origin);
> > -	intel_fbc_flush(dev_priv, frontbuffer_bits, origin);
> > +	intel_edp_drrs_flush(i915, frontbuffer_bits);
> > +	intel_psr_flush(i915, frontbuffer_bits, origin);
> > +	intel_fbc_flush(i915, frontbuffer_bits, origin);
> >   }
> >   void __intel_fb_obj_flush(struct drm_i915_gem_object *obj,
> >   			  enum fb_op_origin origin,
> >   			  unsigned int frontbuffer_bits)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(obj->base.dev);
> >   	if (origin == ORIGIN_CS) {
> > -		spin_lock(&dev_priv->fb_tracking.lock);
> > +		spin_lock(&i915->fb_tracking.lock);
> >   		/* Filter out new bits since rendering started. */
> > -		frontbuffer_bits &= dev_priv->fb_tracking.busy_bits;
> > -		dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > -		spin_unlock(&dev_priv->fb_tracking.lock);
> > +		frontbuffer_bits &= i915->fb_tracking.busy_bits;
> > +		i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > +		spin_unlock(&i915->fb_tracking.lock);
> >   	}
> >   	if (frontbuffer_bits)
> > -		intel_frontbuffer_flush(dev_priv, frontbuffer_bits, origin);
> > +		intel_frontbuffer_flush(i915, frontbuffer_bits, origin);
> >   }
> >   /**
> >    * intel_frontbuffer_flip_prepare - prepare asynchronous frontbuffer flip
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    *
> >    * This function gets called after scheduling a flip on @obj. The actual
> > @@ -141,19 +141,19 @@ void __intel_fb_obj_flush(struct drm_i915_gem_object *obj,
> >    *
> >    * Can be called without any locks held.
> >    */
> > -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915,
> >   				    unsigned frontbuffer_bits)
> >   {
> > -	spin_lock(&dev_priv->fb_tracking.lock);
> > -	dev_priv->fb_tracking.flip_bits |= frontbuffer_bits;
> > +	spin_lock(&i915->fb_tracking.lock);
> > +	i915->fb_tracking.flip_bits |= frontbuffer_bits;
> >   	/* Remove stale busy bits due to the old buffer. */
> > -	dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > -	spin_unlock(&dev_priv->fb_tracking.lock);
> > +	i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > +	spin_unlock(&i915->fb_tracking.lock);
> >   }
> >   /**
> >    * intel_frontbuffer_flip_complete - complete asynchronous frontbuffer flip
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    *
> >    * This function gets called after the flip has been latched and will complete
> > @@ -161,23 +161,23 @@ void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> >    *
> >    * Can be called without any locks held.
> >    */
> > -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915,
> >   				     unsigned frontbuffer_bits)
> >   {
> > -	spin_lock(&dev_priv->fb_tracking.lock);
> > +	spin_lock(&i915->fb_tracking.lock);
> >   	/* Mask any cancelled flips. */
> > -	frontbuffer_bits &= dev_priv->fb_tracking.flip_bits;
> > -	dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits;
> > -	spin_unlock(&dev_priv->fb_tracking.lock);
> > +	frontbuffer_bits &= i915->fb_tracking.flip_bits;
> > +	i915->fb_tracking.flip_bits &= ~frontbuffer_bits;
> > +	spin_unlock(&i915->fb_tracking.lock);
> >   	if (frontbuffer_bits)
> > -		intel_frontbuffer_flush(dev_priv,
> > +		intel_frontbuffer_flush(i915,
> >   					frontbuffer_bits, ORIGIN_FLIP);
> >   }
> >   /**
> >    * intel_frontbuffer_flip - synchronous frontbuffer flip
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    *
> >    * This function gets called after scheduling a flip on @obj. This is for
> > @@ -186,13 +186,13 @@ void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> >    *
> >    * Can be called without any locks held.
> >    */
> > -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip(struct drm_i915_private *i915,
> >   			    unsigned frontbuffer_bits)
> >   {
> > -	spin_lock(&dev_priv->fb_tracking.lock);
> > +	spin_lock(&i915->fb_tracking.lock);
> >   	/* Remove stale busy bits due to the old buffer. */
> > -	dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > -	spin_unlock(&dev_priv->fb_tracking.lock);
> > +	i915->fb_tracking.busy_bits &= ~frontbuffer_bits;
> > +	spin_unlock(&i915->fb_tracking.lock);
> > -	intel_frontbuffer_flush(dev_priv, frontbuffer_bits, ORIGIN_FLIP);
> > +	intel_frontbuffer_flush(i915, frontbuffer_bits, ORIGIN_FLIP);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.h b/drivers/gpu/drm/i915/intel_frontbuffer.h
> > index 5727320c8084..55c42efee7ce 100644
> > --- a/drivers/gpu/drm/i915/intel_frontbuffer.h
> > +++ b/drivers/gpu/drm/i915/intel_frontbuffer.h
> > @@ -37,11 +37,11 @@ enum fb_op_origin {
> >   	ORIGIN_DIRTYFB,
> >   };
> > -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915,
> >   				    unsigned frontbuffer_bits);
> > -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915,
> >   				     unsigned frontbuffer_bits);
> > -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv,
> > +void intel_frontbuffer_flip(struct drm_i915_private *i915,
> >   			    unsigned frontbuffer_bits);
> >   void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj,
> > diff --git a/drivers/gpu/drm/i915/intel_gmbus.c b/drivers/gpu/drm/i915/intel_gmbus.c
> > index aa88e6e7cc65..f1b9844b591e 100644
> > --- a/drivers/gpu/drm/i915/intel_gmbus.c
> > +++ b/drivers/gpu/drm/i915/intel_gmbus.c
> > @@ -89,42 +89,42 @@ static const struct gmbus_pin gmbus_pins_icp[] = {
> >   };
> >   /* pin is expected to be valid */
> > -static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *dev_priv,
> > +static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *i915,
> >   					     unsigned int pin)
> >   {
> > -	if (HAS_PCH_ICP(dev_priv))
> > +	if (HAS_PCH_ICP(i915))
> >   		return &gmbus_pins_icp[pin];
> > -	else if (HAS_PCH_CNP(dev_priv))
> > +	else if (HAS_PCH_CNP(i915))
> >   		return &gmbus_pins_cnp[pin];
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		return &gmbus_pins_bxt[pin];
> > -	else if (IS_GEN9_BC(dev_priv))
> > +	else if (IS_GEN9_BC(i915))
> >   		return &gmbus_pins_skl[pin];
> > -	else if (IS_BROADWELL(dev_priv))
> > +	else if (IS_BROADWELL(i915))
> >   		return &gmbus_pins_bdw[pin];
> >   	else
> >   		return &gmbus_pins[pin];
> >   }
> > -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv,
> > +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915,
> >   			      unsigned int pin)
> >   {
> >   	unsigned int size;
> > -	if (HAS_PCH_ICP(dev_priv))
> > +	if (HAS_PCH_ICP(i915))
> >   		size = ARRAY_SIZE(gmbus_pins_icp);
> > -	else if (HAS_PCH_CNP(dev_priv))
> > +	else if (HAS_PCH_CNP(i915))
> >   		size = ARRAY_SIZE(gmbus_pins_cnp);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		size = ARRAY_SIZE(gmbus_pins_bxt);
> > -	else if (IS_GEN9_BC(dev_priv))
> > +	else if (IS_GEN9_BC(i915))
> >   		size = ARRAY_SIZE(gmbus_pins_skl);
> > -	else if (IS_BROADWELL(dev_priv))
> > +	else if (IS_BROADWELL(i915))
> >   		size = ARRAY_SIZE(gmbus_pins_bdw);
> >   	else
> >   		size = ARRAY_SIZE(gmbus_pins);
> > -	return pin < size && get_gmbus_pin(dev_priv, pin)->name;
> > +	return pin < size && get_gmbus_pin(i915, pin)->name;
> >   }
> >   /* Intel GPIO access functions */
> > @@ -138,13 +138,13 @@ to_intel_gmbus(struct i2c_adapter *i2c)
> >   }
> >   void
> > -intel_gmbus_reset(struct drm_i915_private *dev_priv)
> > +intel_gmbus_reset(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GMBUS0, 0);
> >   	I915_WRITE(GMBUS4, 0);
> >   }
> > -static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> > +static void pnv_gmbus_clock_gating(struct drm_i915_private *i915,
> >   				   bool enable)
> >   {
> >   	u32 val;
> > @@ -158,7 +158,7 @@ static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(DSPCLK_GATE_D, val);
> >   }
> > -static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> > +static void pch_gmbus_clock_gating(struct drm_i915_private *i915,
> >   				   bool enable)
> >   {
> >   	u32 val;
> > @@ -171,7 +171,7 @@ static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(SOUTH_DSPCLK_GATE_D, val);
> >   }
> > -static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> > +static void bxt_gmbus_clock_gating(struct drm_i915_private *i915,
> >   				   bool enable)
> >   {
> >   	u32 val;
> > @@ -186,7 +186,7 @@ static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv,
> >   static u32 get_reserved(struct intel_gmbus *bus)
> >   {
> > -	struct drm_i915_private *i915 = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 reserved = 0;
> > @@ -202,7 +202,7 @@ static u32 get_reserved(struct intel_gmbus *bus)
> >   static int get_clock(void *data)
> >   {
> >   	struct intel_gmbus *bus = data;
> > -	struct intel_uncore *uncore = &bus->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &bus->i915->uncore;
> >   	u32 reserved = get_reserved(bus);
> >   	intel_uncore_write_notrace(uncore,
> > @@ -217,7 +217,7 @@ static int get_clock(void *data)
> >   static int get_data(void *data)
> >   {
> >   	struct intel_gmbus *bus = data;
> > -	struct intel_uncore *uncore = &bus->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &bus->i915->uncore;
> >   	u32 reserved = get_reserved(bus);
> >   	intel_uncore_write_notrace(uncore,
> > @@ -232,7 +232,7 @@ static int get_data(void *data)
> >   static void set_clock(void *data, int state_high)
> >   {
> >   	struct intel_gmbus *bus = data;
> > -	struct intel_uncore *uncore = &bus->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &bus->i915->uncore;
> >   	u32 reserved = get_reserved(bus);
> >   	u32 clock_bits;
> > @@ -251,7 +251,7 @@ static void set_clock(void *data, int state_high)
> >   static void set_data(void *data, int state_high)
> >   {
> >   	struct intel_gmbus *bus = data;
> > -	struct intel_uncore *uncore = &bus->dev_priv->uncore;
> > +	struct intel_uncore *uncore = &bus->i915->uncore;
> >   	u32 reserved = get_reserved(bus);
> >   	u32 data_bits;
> > @@ -271,12 +271,12 @@ intel_gpio_pre_xfer(struct i2c_adapter *adapter)
> >   	struct intel_gmbus *bus = container_of(adapter,
> >   					       struct intel_gmbus,
> >   					       adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> > -	intel_gmbus_reset(dev_priv);
> > +	intel_gmbus_reset(i915);
> > -	if (IS_PINEVIEW(dev_priv))
> > -		pnv_gmbus_clock_gating(dev_priv, false);
> > +	if (IS_PINEVIEW(i915))
> > +		pnv_gmbus_clock_gating(i915, false);
> >   	set_data(bus, 1);
> >   	set_clock(bus, 1);
> > @@ -290,24 +290,24 @@ intel_gpio_post_xfer(struct i2c_adapter *adapter)
> >   	struct intel_gmbus *bus = container_of(adapter,
> >   					       struct intel_gmbus,
> >   					       adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	set_data(bus, 1);
> >   	set_clock(bus, 1);
> > -	if (IS_PINEVIEW(dev_priv))
> > -		pnv_gmbus_clock_gating(dev_priv, true);
> > +	if (IS_PINEVIEW(i915))
> > +		pnv_gmbus_clock_gating(i915, true);
> >   }
> >   static void
> >   intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin)
> >   {
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	struct i2c_algo_bit_data *algo;
> >   	algo = &bus->bit_algo;
> > -	bus->gpio_reg = GPIO(get_gmbus_pin(dev_priv, pin)->gpio);
> > +	bus->gpio_reg = GPIO(get_gmbus_pin(i915, pin)->gpio);
> >   	bus->adapter.algo_data = algo;
> >   	algo->setsda = set_data;
> >   	algo->setscl = set_clock;
> > @@ -320,7 +320,7 @@ intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin)
> >   	algo->data = bus;
> >   }
> > -static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> > +static int gmbus_wait(struct drm_i915_private *i915, u32 status, u32 irq_en)
> >   {
> >   	DEFINE_WAIT(wait);
> >   	u32 gmbus2;
> > @@ -330,10 +330,10 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> >   	 * we also need to check for NAKs besides the hw ready/idle signal, we
> >   	 * need to wake up periodically and check that ourselves.
> >   	 */
> > -	if (!HAS_GMBUS_IRQ(dev_priv))
> > +	if (!HAS_GMBUS_IRQ(i915))
> >   		irq_en = 0;
> > -	add_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> > +	add_wait_queue(&i915->gmbus_wait_queue, &wait);
> >   	I915_WRITE_FW(GMBUS4, irq_en);
> >   	status |= GMBUS_SATOER;
> > @@ -342,7 +342,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> >   		ret = wait_for((gmbus2 = I915_READ_FW(GMBUS2)) & status, 50);
> >   	I915_WRITE_FW(GMBUS4, 0);
> > -	remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> > +	remove_wait_queue(&i915->gmbus_wait_queue, &wait);
> >   	if (gmbus2 & GMBUS_SATOER)
> >   		return -ENXIO;
> > @@ -351,7 +351,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en)
> >   }
> >   static int
> > -gmbus_wait_idle(struct drm_i915_private *dev_priv)
> > +gmbus_wait_idle(struct drm_i915_private *i915)
> >   {
> >   	DEFINE_WAIT(wait);
> >   	u32 irq_enable;
> > @@ -359,36 +359,36 @@ gmbus_wait_idle(struct drm_i915_private *dev_priv)
> >   	/* Important: The hw handles only the first bit, so set only one! */
> >   	irq_enable = 0;
> > -	if (HAS_GMBUS_IRQ(dev_priv))
> > +	if (HAS_GMBUS_IRQ(i915))
> >   		irq_enable = GMBUS_IDLE_EN;
> > -	add_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> > +	add_wait_queue(&i915->gmbus_wait_queue, &wait);
> >   	I915_WRITE_FW(GMBUS4, irq_enable);
> > -	ret = intel_wait_for_register_fw(&dev_priv->uncore,
> > +	ret = intel_wait_for_register_fw(&i915->uncore,
> >   					 GMBUS2, GMBUS_ACTIVE, 0,
> >   					 10);
> >   	I915_WRITE_FW(GMBUS4, 0);
> > -	remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait);
> > +	remove_wait_queue(&i915->gmbus_wait_queue, &wait);
> >   	return ret;
> >   }
> >   static inline
> > -unsigned int gmbus_max_xfer_size(struct drm_i915_private *dev_priv)
> > +unsigned int gmbus_max_xfer_size(struct drm_i915_private *i915)
> >   {
> > -	return INTEL_GEN(dev_priv) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX :
> > +	return INTEL_GEN(i915) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX :
> >   	       GMBUS_BYTE_COUNT_MAX;
> >   }
> >   static int
> > -gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> > +gmbus_xfer_read_chunk(struct drm_i915_private *i915,
> >   		      unsigned short addr, u8 *buf, unsigned int len,
> >   		      u32 gmbus0_reg, u32 gmbus1_index)
> >   {
> >   	unsigned int size = len;
> > -	bool burst_read = len > gmbus_max_xfer_size(dev_priv);
> > +	bool burst_read = len > gmbus_max_xfer_size(i915);
> >   	bool extra_byte_added = false;
> >   	if (burst_read) {
> > @@ -414,7 +414,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> >   		int ret;
> >   		u32 val, loop = 0;
> > -		ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> > +		ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> >   		if (ret)
> >   			return ret;
> > @@ -446,7 +446,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv,
> >   #define INTEL_GMBUS_BURST_READ_MAX_LEN		767U
> >   static int
> > -gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> > +gmbus_xfer_read(struct drm_i915_private *i915, struct i2c_msg *msg,
> >   		u32 gmbus0_reg, u32 gmbus1_index)
> >   {
> >   	u8 *buf = msg->buf;
> > @@ -455,12 +455,12 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> >   	int ret;
> >   	do {
> > -		if (HAS_GMBUS_BURST_READ(dev_priv))
> > +		if (HAS_GMBUS_BURST_READ(i915))
> >   			len = min(rx_size, INTEL_GMBUS_BURST_READ_MAX_LEN);
> >   		else
> > -			len = min(rx_size, gmbus_max_xfer_size(dev_priv));
> > +			len = min(rx_size, gmbus_max_xfer_size(i915));
> > -		ret = gmbus_xfer_read_chunk(dev_priv, msg->addr, buf, len,
> > +		ret = gmbus_xfer_read_chunk(i915, msg->addr, buf, len,
> >   					    gmbus0_reg, gmbus1_index);
> >   		if (ret)
> >   			return ret;
> > @@ -473,7 +473,7 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> >   }
> >   static int
> > -gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
> > +gmbus_xfer_write_chunk(struct drm_i915_private *i915,
> >   		       unsigned short addr, u8 *buf, unsigned int len,
> >   		       u32 gmbus1_index)
> >   {
> > @@ -502,7 +502,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
> >   		I915_WRITE_FW(GMBUS3, val);
> > -		ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> > +		ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN);
> >   		if (ret)
> >   			return ret;
> >   	}
> > @@ -511,7 +511,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv,
> >   }
> >   static int
> > -gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> > +gmbus_xfer_write(struct drm_i915_private *i915, struct i2c_msg *msg,
> >   		 u32 gmbus1_index)
> >   {
> >   	u8 *buf = msg->buf;
> > @@ -520,9 +520,9 @@ gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg,
> >   	int ret;
> >   	do {
> > -		len = min(tx_size, gmbus_max_xfer_size(dev_priv));
> > +		len = min(tx_size, gmbus_max_xfer_size(i915));
> > -		ret = gmbus_xfer_write_chunk(dev_priv, msg->addr, buf, len,
> > +		ret = gmbus_xfer_write_chunk(i915, msg->addr, buf, len,
> >   					     gmbus1_index);
> >   		if (ret)
> >   			return ret;
> > @@ -549,7 +549,7 @@ gmbus_is_index_xfer(struct i2c_msg *msgs, int i, int num)
> >   }
> >   static int
> > -gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs,
> > +gmbus_index_xfer(struct drm_i915_private *i915, struct i2c_msg *msgs,
> >   		 u32 gmbus0_reg)
> >   {
> >   	u32 gmbus1_index = 0;
> > @@ -568,10 +568,10 @@ gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs,
> >   		I915_WRITE_FW(GMBUS5, gmbus5);
> >   	if (msgs[1].flags & I2C_M_RD)
> > -		ret = gmbus_xfer_read(dev_priv, &msgs[1], gmbus0_reg,
> > +		ret = gmbus_xfer_read(i915, &msgs[1], gmbus0_reg,
> >   				      gmbus1_index);
> >   	else
> > -		ret = gmbus_xfer_write(dev_priv, &msgs[1], gmbus1_index);
> > +		ret = gmbus_xfer_write(i915, &msgs[1], gmbus1_index);
> >   	/* Clear GMBUS5 after each index transfer */
> >   	if (gmbus5)
> > @@ -587,15 +587,15 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> >   	struct intel_gmbus *bus = container_of(adapter,
> >   					       struct intel_gmbus,
> >   					       adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	int i = 0, inc, try = 0;
> >   	int ret = 0;
> >   	/* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */
> > -	if (IS_GEN9_LP(dev_priv))
> > -		bxt_gmbus_clock_gating(dev_priv, false);
> > -	else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv))
> > -		pch_gmbus_clock_gating(dev_priv, false);
> > +	if (IS_GEN9_LP(i915))
> > +		bxt_gmbus_clock_gating(i915, false);
> > +	else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915))
> > +		pch_gmbus_clock_gating(i915, false);
> >   retry:
> >   	I915_WRITE_FW(GMBUS0, gmbus0_source | bus->reg0);
> > @@ -603,18 +603,18 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> >   	for (; i < num; i += inc) {
> >   		inc = 1;
> >   		if (gmbus_is_index_xfer(msgs, i, num)) {
> > -			ret = gmbus_index_xfer(dev_priv, &msgs[i],
> > +			ret = gmbus_index_xfer(i915, &msgs[i],
> >   					       gmbus0_source | bus->reg0);
> >   			inc = 2; /* an index transmission is two msgs */
> >   		} else if (msgs[i].flags & I2C_M_RD) {
> > -			ret = gmbus_xfer_read(dev_priv, &msgs[i],
> > +			ret = gmbus_xfer_read(i915, &msgs[i],
> >   					      gmbus0_source | bus->reg0, 0);
> >   		} else {
> > -			ret = gmbus_xfer_write(dev_priv, &msgs[i], 0);
> > +			ret = gmbus_xfer_write(i915, &msgs[i], 0);
> >   		}
> >   		if (!ret)
> > -			ret = gmbus_wait(dev_priv,
> > +			ret = gmbus_wait(i915,
> >   					 GMBUS_HW_WAIT_PHASE, GMBUS_HW_WAIT_EN);
> >   		if (ret == -ETIMEDOUT)
> >   			goto timeout;
> > @@ -632,7 +632,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> >   	 * We will re-enable it at the start of the next xfer,
> >   	 * till then let it sleep.
> >   	 */
> > -	if (gmbus_wait_idle(dev_priv)) {
> > +	if (gmbus_wait_idle(i915)) {
> >   		DRM_DEBUG_KMS("GMBUS [%s] timed out waiting for idle\n",
> >   			 adapter->name);
> >   		ret = -ETIMEDOUT;
> > @@ -656,7 +656,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> >   	 * it's slow responding and only answers on the 2nd retry.
> >   	 */
> >   	ret = -ENXIO;
> > -	if (gmbus_wait_idle(dev_priv)) {
> > +	if (gmbus_wait_idle(i915)) {
> >   		DRM_DEBUG_KMS("GMBUS [%s] timed out after NAK\n",
> >   			      adapter->name);
> >   		ret = -ETIMEDOUT;
> > @@ -701,10 +701,10 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num,
> >   out:
> >   	/* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */
> > -	if (IS_GEN9_LP(dev_priv))
> > -		bxt_gmbus_clock_gating(dev_priv, true);
> > -	else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv))
> > -		pch_gmbus_clock_gating(dev_priv, true);
> > +	if (IS_GEN9_LP(i915))
> > +		bxt_gmbus_clock_gating(i915, true);
> > +	else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915))
> > +		pch_gmbus_clock_gating(i915, true);
> >   	return ret;
> >   }
> > @@ -714,11 +714,11 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num)
> >   {
> >   	struct intel_gmbus *bus =
> >   		container_of(adapter, struct intel_gmbus, adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	intel_wakeref_t wakeref;
> >   	int ret;
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
> >   	if (bus->force_bit) {
> >   		ret = i2c_bit_algo.master_xfer(adapter, msgs, num);
> > @@ -730,7 +730,7 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num)
> >   			bus->force_bit |= GMBUS_FORCE_BIT_RETRY;
> >   	}
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
> >   	return ret;
> >   }
> > @@ -739,7 +739,7 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> >   {
> >   	struct intel_gmbus *bus =
> >   		container_of(adapter, struct intel_gmbus, adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> >   	u8 cmd = DRM_HDCP_DDC_AKSV;
> >   	u8 buf[DRM_HDCP_KSV_LEN] = { 0 };
> >   	struct i2c_msg msgs[] = {
> > @@ -759,8 +759,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> >   	intel_wakeref_t wakeref;
> >   	int ret;
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> > -	mutex_lock(&dev_priv->gmbus_mutex);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
> > +	mutex_lock(&i915->gmbus_mutex);
> >   	/*
> >   	 * In order to output Aksv to the receiver, use an indexed write to
> > @@ -769,8 +769,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter)
> >   	 */
> >   	ret = do_gmbus_xfer(adapter, msgs, ARRAY_SIZE(msgs), GMBUS_AKSV_SELECT);
> > -	mutex_unlock(&dev_priv->gmbus_mutex);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> > +	mutex_unlock(&i915->gmbus_mutex);
> > +	intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
> >   	return ret;
> >   }
> > @@ -793,27 +793,27 @@ static void gmbus_lock_bus(struct i2c_adapter *adapter,
> >   			   unsigned int flags)
> >   {
> >   	struct intel_gmbus *bus = to_intel_gmbus(adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> > -	mutex_lock(&dev_priv->gmbus_mutex);
> > +	mutex_lock(&i915->gmbus_mutex);
> >   }
> >   static int gmbus_trylock_bus(struct i2c_adapter *adapter,
> >   			     unsigned int flags)
> >   {
> >   	struct intel_gmbus *bus = to_intel_gmbus(adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> > -	return mutex_trylock(&dev_priv->gmbus_mutex);
> > +	return mutex_trylock(&i915->gmbus_mutex);
> >   }
> >   static void gmbus_unlock_bus(struct i2c_adapter *adapter,
> >   			     unsigned int flags)
> >   {
> >   	struct intel_gmbus *bus = to_intel_gmbus(adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> > -	mutex_unlock(&dev_priv->gmbus_mutex);
> > +	mutex_unlock(&i915->gmbus_mutex);
> >   }
> >   static const struct i2c_lock_operations gmbus_lock_ops = {
> > @@ -824,45 +824,45 @@ static const struct i2c_lock_operations gmbus_lock_ops = {
> >   /**
> >    * intel_gmbus_setup - instantiate all Intel i2c GMBuses
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    */
> > -int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> > +int intel_gmbus_setup(struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	struct intel_gmbus *bus;
> >   	unsigned int pin;
> >   	int ret;
> > -	if (!HAS_DISPLAY(dev_priv))
> > +	if (!HAS_DISPLAY(i915))
> >   		return 0;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->gpio_mmio_base = VLV_DISPLAY_BASE;
> > -	else if (!HAS_GMCH(dev_priv))
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->gpio_mmio_base = VLV_DISPLAY_BASE;
> > +	else if (!HAS_GMCH(i915))
> >   		/*
> >   		 * Broxton uses the same PCH offsets for South Display Engine,
> >   		 * even though it doesn't have a PCH.
> >   		 */
> > -		dev_priv->gpio_mmio_base = PCH_DISPLAY_BASE;
> > +		i915->gpio_mmio_base = PCH_DISPLAY_BASE;
> > -	mutex_init(&dev_priv->gmbus_mutex);
> > -	init_waitqueue_head(&dev_priv->gmbus_wait_queue);
> > +	mutex_init(&i915->gmbus_mutex);
> > +	init_waitqueue_head(&i915->gmbus_wait_queue);
> > -	for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) {
> > -		if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> > +	for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) {
> > +		if (!intel_gmbus_is_valid_pin(i915, pin))
> >   			continue;
> > -		bus = &dev_priv->gmbus[pin];
> > +		bus = &i915->gmbus[pin];
> >   		bus->adapter.owner = THIS_MODULE;
> >   		bus->adapter.class = I2C_CLASS_DDC;
> >   		snprintf(bus->adapter.name,
> >   			 sizeof(bus->adapter.name),
> >   			 "i915 gmbus %s",
> > -			 get_gmbus_pin(dev_priv, pin)->name);
> > +			 get_gmbus_pin(i915, pin)->name);
> >   		bus->adapter.dev.parent = &pdev->dev;
> > -		bus->dev_priv = dev_priv;
> > +		bus->i915 = i915;
> >   		bus->adapter.algo = &gmbus_algorithm;
> >   		bus->adapter.lock_ops = &gmbus_lock_ops;
> > @@ -877,7 +877,7 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> >   		bus->reg0 = pin | GMBUS_RATE_100KHZ;
> >   		/* gmbus seems to be broken on i830 */
> > -		if (IS_I830(dev_priv))
> > +		if (IS_I830(i915))
> >   			bus->force_bit = 1;
> >   		intel_gpio_setup(bus, pin);
> > @@ -887,28 +887,28 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv)
> >   			goto err;
> >   	}
> > -	intel_gmbus_reset(dev_priv);
> > +	intel_gmbus_reset(i915);
> >   	return 0;
> >   err:
> >   	while (pin--) {
> > -		if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> > +		if (!intel_gmbus_is_valid_pin(i915, pin))
> >   			continue;
> > -		bus = &dev_priv->gmbus[pin];
> > +		bus = &i915->gmbus[pin];
> >   		i2c_del_adapter(&bus->adapter);
> >   	}
> >   	return ret;
> >   }
> > -struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *dev_priv,
> > +struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *i915,
> >   					    unsigned int pin)
> >   {
> > -	if (WARN_ON(!intel_gmbus_is_valid_pin(dev_priv, pin)))
> > +	if (WARN_ON(!intel_gmbus_is_valid_pin(i915, pin)))
> >   		return NULL;
> > -	return &dev_priv->gmbus[pin].adapter;
> > +	return &i915->gmbus[pin].adapter;
> >   }
> >   void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed)
> > @@ -921,16 +921,16 @@ void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed)
> >   void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit)
> >   {
> >   	struct intel_gmbus *bus = to_intel_gmbus(adapter);
> > -	struct drm_i915_private *dev_priv = bus->dev_priv;
> > +	struct drm_i915_private *i915 = bus->i915;
> > -	mutex_lock(&dev_priv->gmbus_mutex);
> > +	mutex_lock(&i915->gmbus_mutex);
> >   	bus->force_bit += force_bit ? 1 : -1;
> >   	DRM_DEBUG_KMS("%sabling bit-banging on %s. force bit now %d\n",
> >   		      force_bit ? "en" : "dis", adapter->name,
> >   		      bus->force_bit);
> > -	mutex_unlock(&dev_priv->gmbus_mutex);
> > +	mutex_unlock(&i915->gmbus_mutex);
> >   }
> >   bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter)
> > @@ -940,16 +940,16 @@ bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter)
> >   	return bus->force_bit;
> >   }
> > -void intel_gmbus_teardown(struct drm_i915_private *dev_priv)
> > +void intel_gmbus_teardown(struct drm_i915_private *i915)
> >   {
> >   	struct intel_gmbus *bus;
> >   	unsigned int pin;
> > -	for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) {
> > -		if (!intel_gmbus_is_valid_pin(dev_priv, pin))
> > +	for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) {
> > +		if (!intel_gmbus_is_valid_pin(i915, pin))
> >   			continue;
> > -		bus = &dev_priv->gmbus[pin];
> > +		bus = &i915->gmbus[pin];
> >   		i2c_del_adapter(&bus->adapter);
> >   	}
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_gmbus.h b/drivers/gpu/drm/i915/intel_gmbus.h
> > index d989085b8d22..4a4196ca8343 100644
> > --- a/drivers/gpu/drm/i915/intel_gmbus.h
> > +++ b/drivers/gpu/drm/i915/intel_gmbus.h
> > @@ -11,17 +11,17 @@
> >   struct drm_i915_private;
> >   struct i2c_adapter;
> > -int intel_gmbus_setup(struct drm_i915_private *dev_priv);
> > -void intel_gmbus_teardown(struct drm_i915_private *dev_priv);
> > -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv,
> > +int intel_gmbus_setup(struct drm_i915_private *i915);
> > +void intel_gmbus_teardown(struct drm_i915_private *i915);
> > +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915,
> >   			      unsigned int pin);
> >   int intel_gmbus_output_aksv(struct i2c_adapter *adapter);
> >   struct i2c_adapter *
> > -intel_gmbus_get_adapter(struct drm_i915_private *dev_priv, unsigned int pin);
> > +intel_gmbus_get_adapter(struct drm_i915_private *i915, unsigned int pin);
> >   void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed);
> >   void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit);
> >   bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter);
> > -void intel_gmbus_reset(struct drm_i915_private *dev_priv);
> > +void intel_gmbus_reset(struct drm_i915_private *i915);
> >   #endif /* __INTEL_GMBUS_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_guc.c b/drivers/gpu/drm/i915/intel_guc.c
> > index c40a6efdd33a..f08ac29c2a4f 100644
> > --- a/drivers/gpu/drm/i915/intel_guc.c
> > +++ b/drivers/gpu/drm/i915/intel_guc.c
> > @@ -29,14 +29,14 @@
> >   static void gen8_guc_raise_irq(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	I915_WRITE(GUC_SEND_INTERRUPT, GUC_SEND_TRIGGER);
> >   }
> >   static void gen11_guc_raise_irq(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	I915_WRITE(GEN11_GUC_HOST_INTERRUPT, 0);
> >   }
> > @@ -52,11 +52,11 @@ static inline i915_reg_t guc_send_reg(struct intel_guc *guc, u32 i)
> >   void intel_guc_init_send_regs(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	enum forcewake_domains fw_domains = 0;
> >   	unsigned int i;
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		guc->send_regs.base =
> >   				i915_mmio_reg_offset(GEN11_SOFT_SCRATCH(0));
> >   		guc->send_regs.count = GEN11_SOFT_SCRATCH_COUNT;
> > @@ -67,7 +67,7 @@ void intel_guc_init_send_regs(struct intel_guc *guc)
> >   	}
> >   	for (i = 0; i < guc->send_regs.count; i++) {
> > -		fw_domains |= intel_uncore_forcewake_for_reg(&dev_priv->uncore,
> > +		fw_domains |= intel_uncore_forcewake_for_reg(&i915->uncore,
> >   					guc_send_reg(guc, i),
> >   					FW_REG_READ | FW_REG_WRITE);
> >   	}
> > @@ -101,7 +101,7 @@ void intel_guc_init_early(struct intel_guc *guc)
> >   static int guc_init_wq(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	/*
> >   	 * GuC log buffer flush work item has to do register access to
> > @@ -135,8 +135,8 @@ static int guc_init_wq(struct intel_guc *guc)
> >   	 * to make sure we're always sending a single preemption request with a
> >   	 * single workitem.
> >   	 */
> > -	if (HAS_LOGICAL_RING_PREEMPTION(dev_priv) &&
> > -	    USES_GUC_SUBMISSION(dev_priv)) {
> > +	if (HAS_LOGICAL_RING_PREEMPTION(i915) &&
> > +	    USES_GUC_SUBMISSION(i915)) {
> >   		guc->preempt_wq = alloc_ordered_workqueue("i915-guc_preempt",
> >   							  WQ_HIGHPRI);
> >   		if (!guc->preempt_wq) {
> > @@ -211,7 +211,7 @@ static void guc_shared_data_destroy(struct intel_guc *guc)
> >   int intel_guc_init(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	int ret;
> >   	ret = intel_uc_fw_init(&guc->fw);
> > @@ -237,7 +237,7 @@ int intel_guc_init(struct intel_guc *guc)
> >   		goto err_ads;
> >   	/* We need to notify the guc whenever we change the GGTT */
> > -	i915_ggtt_enable_guc(dev_priv);
> > +	i915_ggtt_enable_guc(i915);
> >   	return 0;
> > @@ -256,9 +256,9 @@ int intel_guc_init(struct intel_guc *guc)
> >   void intel_guc_fini(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> > -	i915_ggtt_disable_guc(dev_priv);
> > +	i915_ggtt_disable_guc(i915);
> >   	intel_guc_ct_fini(&guc->ct);
> > @@ -366,7 +366,7 @@ static u32 guc_ctl_ads_flags(struct intel_guc *guc)
> >    */
> >   void intel_guc_init_params(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	u32 params[GUC_CTL_MAX_DWORDS];
> >   	int i;
> > @@ -386,14 +386,14 @@ void intel_guc_init_params(struct intel_guc *guc)
> >   	 * they are power context saved so it's ok to release forcewake
> >   	 * when we are done here and take it again at xfer time.
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_BLITTER);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_BLITTER);
> >   	I915_WRITE(SOFT_SCRATCH(0), 0);
> >   	for (i = 0; i < GUC_CTL_MAX_DWORDS; i++)
> >   		I915_WRITE(SOFT_SCRATCH(1 + i), params[i]);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_BLITTER);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_BLITTER);
> >   }
> >   int intel_guc_send_nop(struct intel_guc *guc, const u32 *action, u32 len,
> > @@ -414,8 +414,8 @@ void intel_guc_to_host_event_handler_nop(struct intel_guc *guc)
> >   int intel_guc_send_mmio(struct intel_guc *guc, const u32 *action, u32 len,
> >   			u32 *response_buf, u32 response_buf_size)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 status;
> >   	int i;
> >   	int ret;
> > @@ -497,12 +497,12 @@ int intel_guc_to_host_process_recv_msg(struct intel_guc *guc,
> >   int intel_guc_sample_forcewake(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	u32 action[2];
> >   	action[0] = INTEL_GUC_ACTION_SAMPLE_FORCEWAKE;
> >   	/* WaRsDisableCoarsePowerGating:skl,cnl */
> > -	if (!HAS_RC6(dev_priv) || NEEDS_WaRsDisableCoarsePowerGating(dev_priv))
> > +	if (!HAS_RC6(i915) || NEEDS_WaRsDisableCoarsePowerGating(i915))
> >   		action[1] = 0;
> >   	else
> >   		/* bit 0 and 1 are for Render and Media domain separately */
> > @@ -538,7 +538,7 @@ int intel_guc_auth_huc(struct intel_guc *guc, u32 rsa_offset)
> >    */
> >   int intel_guc_suspend(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	int ret;
> >   	u32 status;
> >   	u32 action[] = {
> > @@ -562,7 +562,7 @@ int intel_guc_suspend(struct intel_guc *guc)
> >   	if (ret)
> >   		return ret;
> > -	ret = __intel_wait_for_register(&dev_priv->uncore, SOFT_SCRATCH(14),
> > +	ret = __intel_wait_for_register(&i915->uncore, SOFT_SCRATCH(14),
> >   					INTEL_GUC_SLEEP_STATE_INVALID_MASK,
> >   					0, 0, 10, &status);
> >   	if (ret)
> > @@ -658,17 +658,17 @@ int intel_guc_resume(struct intel_guc *guc)
> >    */
> >   struct i915_vma *intel_guc_allocate_vma(struct intel_guc *guc, u32 size)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct drm_i915_gem_object *obj;
> >   	struct i915_vma *vma;
> >   	u64 flags;
> >   	int ret;
> > -	obj = i915_gem_object_create_shmem(dev_priv, size);
> > +	obj = i915_gem_object_create_shmem(i915, size);
> >   	if (IS_ERR(obj))
> >   		return ERR_CAST(obj);
> > -	vma = i915_vma_instance(obj, &dev_priv->ggtt.vm, NULL);
> > +	vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
> >   	if (IS_ERR(vma))
> >   		goto err;
> > diff --git a/drivers/gpu/drm/i915/intel_guc_ads.c b/drivers/gpu/drm/i915/intel_guc_ads.c
> > index ecb69fc94218..4ac9ed2dd467 100644
> > --- a/drivers/gpu/drm/i915/intel_guc_ads.c
> > +++ b/drivers/gpu/drm/i915/intel_guc_ads.c
> > @@ -85,7 +85,7 @@ struct __guc_ads_blob {
> >   static int __guc_ads_init(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct __guc_ads_blob *blob;
> >   	const u32 skipped_size = LRC_PPHWSP_SZ * PAGE_SIZE + LR_HW_CONTEXT_SIZE;
> >   	u32 base;
> > @@ -115,18 +115,18 @@ static int __guc_ads_init(struct intel_guc *guc)
> >   		 */
> >   		blob->ads.golden_context_lrca[engine_class] = 0;
> >   		blob->ads.eng_state_size[engine_class] =
> > -			intel_engine_context_size(dev_priv, engine_class) -
> > +			intel_engine_context_size(i915, engine_class) -
> >   			skipped_size;
> >   	}
> >   	/* System info */
> > -	blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(dev_priv)->sseu.slice_mask);
> > +	blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(i915)->sseu.slice_mask);
> >   	blob->system_info.rcs_enabled = 1;
> >   	blob->system_info.bcs_enabled = 1;
> > -	blob->system_info.vdbox_enable_mask = VDBOX_MASK(dev_priv);
> > -	blob->system_info.vebox_enable_mask = VEBOX_MASK(dev_priv);
> > -	blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(dev_priv)->vdbox_sfc_access;
> > +	blob->system_info.vdbox_enable_mask = VDBOX_MASK(i915);
> > +	blob->system_info.vebox_enable_mask = VEBOX_MASK(i915);
> > +	blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(i915)->vdbox_sfc_access;
> >   	base = intel_guc_ggtt_offset(guc, guc->ads_vma);
> > diff --git a/drivers/gpu/drm/i915/intel_guc_fw.c b/drivers/gpu/drm/i915/intel_guc_fw.c
> > index 72cdafd9636a..2beb90c62a40 100644
> > --- a/drivers/gpu/drm/i915/intel_guc_fw.c
> > +++ b/drivers/gpu/drm/i915/intel_guc_fw.c
> > @@ -125,7 +125,7 @@ void intel_guc_fw_init_early(struct intel_guc *guc)
> >   static void guc_prepare_xfer(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	/* Must program this register before loading the ucode with DMA */
> >   	I915_WRITE(GUC_SHIM_CONTROL, GUC_DISABLE_SRAM_INIT_TO_ZEROES |
> > @@ -135,12 +135,12 @@ static void guc_prepare_xfer(struct intel_guc *guc)
> >   				     GUC_ENABLE_READ_CACHE_FOR_WOPCM_DATA |
> >   				     GUC_ENABLE_MIA_CLOCK_GATING);
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		I915_WRITE(GEN9LP_GT_PM_CONFIG, GT_DOORBELL_ENABLE);
> >   	else
> >   		I915_WRITE(GEN9_GT_PM_CONFIG, GT_DOORBELL_ENABLE);
> > -	if (IS_GEN(dev_priv, 9)) {
> > +	if (IS_GEN(i915, 9)) {
> >   		/* DOP Clock Gating Enable for GuC clocks */
> >   		I915_WRITE(GEN7_MISCCPCTL, (GEN8_DOP_CLOCK_GATE_GUC_ENABLE |
> >   					    I915_READ(GEN7_MISCCPCTL)));
> > @@ -153,7 +153,7 @@ static void guc_prepare_xfer(struct intel_guc *guc)
> >   /* Copy RSA signature from the fw image to HW for verification */
> >   static void guc_xfer_rsa(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_uc_fw *fw = &guc->fw;
> >   	struct sg_table *pages = fw->obj->mm.pages;
> >   	u32 rsa[UOS_RSA_SCRATCH_COUNT];
> > @@ -168,7 +168,7 @@ static void guc_xfer_rsa(struct intel_guc *guc)
> >   static bool guc_xfer_completed(struct intel_guc *guc, u32 *status)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	/* Did we complete the xfer? */
> >   	*status = I915_READ(DMA_CTRL);
> > @@ -186,7 +186,7 @@ static bool guc_xfer_completed(struct intel_guc *guc, u32 *status)
> >    */
> >   static inline bool guc_ready(struct intel_guc *guc, u32 *status)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	u32 val = I915_READ(GUC_STATUS);
> >   	u32 uk_val = val & GS_UKERNEL_MASK;
> > @@ -234,7 +234,7 @@ static int guc_wait_ucode(struct intel_guc *guc)
> >    */
> >   static int guc_xfer_ucode(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_uc_fw *guc_fw = &guc->fw;
> >   	unsigned long offset;
> > @@ -267,12 +267,12 @@ static int guc_xfer_ucode(struct intel_guc *guc)
> >   static int guc_fw_xfer(struct intel_uc_fw *guc_fw)
> >   {
> >   	struct intel_guc *guc = container_of(guc_fw, struct intel_guc, fw);
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	int ret;
> >   	GEM_BUG_ON(guc_fw->type != INTEL_UC_FW_TYPE_GUC);
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	guc_prepare_xfer(guc);
> > @@ -285,7 +285,7 @@ static int guc_fw_xfer(struct intel_uc_fw *guc_fw)
> >   	ret = guc_xfer_ucode(guc);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   	return ret;
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_guc_log.c b/drivers/gpu/drm/i915/intel_guc_log.c
> > index 67eadc82c396..300ed940f0b7 100644
> > --- a/drivers/gpu/drm/i915/intel_guc_log.c
> > +++ b/drivers/gpu/drm/i915/intel_guc_log.c
> > @@ -390,7 +390,7 @@ void intel_guc_log_init_early(struct intel_guc_log *log)
> >   static int guc_log_relay_create(struct intel_guc_log *log)
> >   {
> >   	struct intel_guc *guc = log_to_guc(log);
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct rchan *guc_log_relay_chan;
> >   	size_t n_subbufs, subbuf_size;
> >   	int ret;
> > @@ -409,9 +409,9 @@ static int guc_log_relay_create(struct intel_guc_log *log)
> >   	n_subbufs = 8;
> >   	guc_log_relay_chan = relay_open("guc_log",
> > -					dev_priv->drm.primary->debugfs_root,
> > +					i915->drm.primary->debugfs_root,
> >   					subbuf_size, n_subbufs,
> > -					&relay_callbacks, dev_priv);
> > +					&relay_callbacks, i915);
> >   	if (!guc_log_relay_chan) {
> >   		DRM_ERROR("Couldn't create relay chan for GuC logging\n");
> > @@ -436,7 +436,7 @@ static void guc_log_relay_destroy(struct intel_guc_log *log)
> >   static void guc_log_capture_logs(struct intel_guc_log *log)
> >   {
> >   	struct intel_guc *guc = log_to_guc(log);
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	intel_wakeref_t wakeref;
> >   	guc_read_update_log_buffer(log);
> > @@ -445,7 +445,7 @@ static void guc_log_capture_logs(struct intel_guc_log *log)
> >   	 * Generally device is expected to be active only at this
> >   	 * time, so get/put should be really quick.
> >   	 */
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		guc_action_flush_log_complete(guc);
> >   }
> > @@ -505,7 +505,7 @@ void intel_guc_log_destroy(struct intel_guc_log *log)
> >   int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> >   {
> >   	struct intel_guc *guc = log_to_guc(log);
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	intel_wakeref_t wakeref;
> >   	int ret = 0;
> > @@ -519,12 +519,12 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> >   	if (level < GUC_LOG_LEVEL_DISABLED || level > GUC_LOG_LEVEL_MAX)
> >   		return -EINVAL;
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	if (log->level == level)
> >   		goto out_unlock;
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		ret = guc_action_control_log(guc,
> >   					     GUC_LOG_LEVEL_IS_VERBOSE(level),
> >   					     GUC_LOG_LEVEL_IS_ENABLED(level),
> > @@ -537,7 +537,7 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level)
> >   	log->level = level;
> >   out_unlock:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return ret;
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_guc_submission.c b/drivers/gpu/drm/i915/intel_guc_submission.c
> > index 97f6970d8da8..013d85278270 100644
> > --- a/drivers/gpu/drm/i915/intel_guc_submission.c
> > +++ b/drivers/gpu/drm/i915/intel_guc_submission.c
> > @@ -203,7 +203,7 @@ static struct guc_doorbell_info *__get_doorbell(struct intel_guc_client *client)
> >   static bool __doorbell_valid(struct intel_guc *guc, u16 db_id)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	GEM_BUG_ON(db_id >= GUC_NUM_DOORBELLS);
> >   	return I915_READ(GEN8_DRBREGL(db_id)) & GEN8_DRB_VALID;
> > @@ -1014,7 +1014,7 @@ static bool guc_verify_doorbells(struct intel_guc *guc)
> >   /**
> >    * guc_client_alloc() - Allocate an intel_guc_client
> > - * @dev_priv:	driver private data structure
> > + * @i915:	driver private data structure
> >    * @engines:	The set of engines to enable for this client
> >    * @priority:	four levels priority _CRITICAL, _HIGH, _NORMAL and _LOW
> >    *		The kernel client to replace ExecList submission is created with
> > @@ -1026,13 +1026,13 @@ static bool guc_verify_doorbells(struct intel_guc *guc)
> >    * Return:	An intel_guc_client object if success, else NULL.
> >    */
> >   static struct intel_guc_client *
> > -guc_client_alloc(struct drm_i915_private *dev_priv,
> > +guc_client_alloc(struct drm_i915_private *i915,
> >   		 u32 engines,
> >   		 u32 priority,
> >   		 struct i915_gem_context *ctx)
> >   {
> >   	struct intel_guc_client *client;
> > -	struct intel_guc *guc = &dev_priv->guc;
> > +	struct intel_guc *guc = &i915->guc;
> >   	struct i915_vma *vma;
> >   	void *vaddr;
> >   	int ret;
> > @@ -1129,27 +1129,27 @@ static inline bool ctx_save_restore_disabled(struct intel_context *ce)
> >   static int guc_clients_create(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_guc_client *client;
> >   	GEM_BUG_ON(guc->execbuf_client);
> >   	GEM_BUG_ON(guc->preempt_client);
> > -	client = guc_client_alloc(dev_priv,
> > -				  INTEL_INFO(dev_priv)->engine_mask,
> > +	client = guc_client_alloc(i915,
> > +				  INTEL_INFO(i915)->engine_mask,
> >   				  GUC_CLIENT_PRIORITY_KMD_NORMAL,
> > -				  dev_priv->kernel_context);
> > +				  i915->kernel_context);
> >   	if (IS_ERR(client)) {
> >   		DRM_ERROR("Failed to create GuC client for submission!\n");
> >   		return PTR_ERR(client);
> >   	}
> >   	guc->execbuf_client = client;
> > -	if (dev_priv->preempt_context) {
> > -		client = guc_client_alloc(dev_priv,
> > -					  INTEL_INFO(dev_priv)->engine_mask,
> > +	if (i915->preempt_context) {
> > +		client = guc_client_alloc(i915,
> > +					  INTEL_INFO(i915)->engine_mask,
> >   					  GUC_CLIENT_PRIORITY_KMD_HIGH,
> > -					  dev_priv->preempt_context);
> > +					  i915->preempt_context);
> >   		if (IS_ERR(client)) {
> >   			DRM_ERROR("Failed to create GuC client for preemption!\n");
> >   			guc_client_free(guc->execbuf_client);
> > @@ -1244,7 +1244,7 @@ static void guc_clients_disable(struct intel_guc *guc)
> >    */
> >   int intel_guc_submission_init(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	int ret;
> > @@ -1266,7 +1266,7 @@ int intel_guc_submission_init(struct intel_guc *guc)
> >   	if (ret)
> >   		goto err_pool;
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		guc->preempt_work[id].engine = engine;
> >   		INIT_WORK(&guc->preempt_work[id].work, inject_preempt_context);
> >   	}
> > @@ -1280,11 +1280,11 @@ int intel_guc_submission_init(struct intel_guc *guc)
> >   void intel_guc_submission_fini(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		cancel_work_sync(&guc->preempt_work[id].work);
> >   	guc_clients_destroy(guc);
> > @@ -1294,9 +1294,9 @@ void intel_guc_submission_fini(struct intel_guc *guc)
> >   		guc_stage_desc_pool_destroy(guc);
> >   }
> > -static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> > +static void guc_interrupts_capture(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	int irqs;
> > @@ -1305,7 +1305,7 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> >   	 * to GuC
> >   	 */
> >   	irqs = _MASKED_BIT_ENABLE(GFX_INTERRUPT_STEERING);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		ENGINE_WRITE(engine, RING_MODE_GEN7, irqs);
> >   	/* route USER_INTERRUPT to Host, all others are sent to GuC. */
> > @@ -1339,9 +1339,9 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
> >   	rps->pm_intrmsk_mbz &= ~GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC;
> >   }
> > -static void guc_interrupts_release(struct drm_i915_private *dev_priv)
> > +static void guc_interrupts_release(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	int irqs;
> > @@ -1352,7 +1352,7 @@ static void guc_interrupts_release(struct drm_i915_private *dev_priv)
> >   	 */
> >   	irqs = _MASKED_FIELD(GFX_FORWARD_VBLANK_MASK, GFX_FORWARD_VBLANK_NEVER);
> >   	irqs |= _MASKED_BIT_DISABLE(GFX_INTERRUPT_STEERING);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		ENGINE_WRITE(engine, RING_MODE_GEN7, irqs);
> >   	/* route all GT interrupts to the host */
> > @@ -1408,7 +1408,7 @@ static void guc_set_default_submission(struct intel_engine_cs *engine)
> >   int intel_guc_submission_enable(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	int err;
> > @@ -1433,9 +1433,9 @@ int intel_guc_submission_enable(struct intel_guc *guc)
> >   		return err;
> >   	/* Take over from manual control of ELSP (execlists) */
> > -	guc_interrupts_capture(dev_priv);
> > +	guc_interrupts_capture(i915);
> > -	for_each_engine(engine, dev_priv, id) {
> > +	for_each_engine(engine, i915, id) {
> >   		engine->set_default_submission = guc_set_default_submission;
> >   		engine->set_default_submission(engine);
> >   	}
> > @@ -1445,11 +1445,11 @@ int intel_guc_submission_enable(struct intel_guc *guc)
> >   void intel_guc_submission_disable(struct intel_guc *guc)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(guc);
> > -	GEM_BUG_ON(dev_priv->gt.awake); /* GT should be parked first */
> > +	GEM_BUG_ON(i915->gt.awake); /* GT should be parked first */
> > -	guc_interrupts_release(dev_priv);
> > +	guc_interrupts_release(i915);
> >   	guc_clients_disable(guc);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_gvt.c b/drivers/gpu/drm/i915/intel_gvt.c
> > index 1d7d26e4cf14..b50d31a2e095 100644
> > --- a/drivers/gpu/drm/i915/intel_gvt.c
> > +++ b/drivers/gpu/drm/i915/intel_gvt.c
> > @@ -39,17 +39,17 @@
> >    * doc is available on https://01.org/group/2230/documentation-list.
> >    */
> > -static bool is_supported_device(struct drm_i915_private *dev_priv)
> > +static bool is_supported_device(struct drm_i915_private *i915)
> >   {
> > -	if (IS_BROADWELL(dev_priv))
> > +	if (IS_BROADWELL(i915))
> >   		return true;
> > -	if (IS_SKYLAKE(dev_priv))
> > +	if (IS_SKYLAKE(i915))
> >   		return true;
> > -	if (IS_KABYLAKE(dev_priv))
> > +	if (IS_KABYLAKE(i915))
> >   		return true;
> > -	if (IS_BROXTON(dev_priv))
> > +	if (IS_BROXTON(i915))
> >   		return true;
> > -	if (IS_COFFEELAKE(dev_priv))
> > +	if (IS_COFFEELAKE(i915))
> >   		return true;
> >   	return false;
> > @@ -57,21 +57,21 @@ static bool is_supported_device(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_gvt_sanitize_options - sanitize GVT related options
> > - * @dev_priv: drm i915 private data
> > + * @i915: drm i915 private data
> >    *
> >    * This function is called at the i915 options sanitize stage.
> >    */
> > -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> > +void intel_gvt_sanitize_options(struct drm_i915_private *i915)
> >   {
> >   	if (!i915_modparams.enable_gvt)
> >   		return;
> > -	if (intel_vgpu_active(dev_priv)) {
> > +	if (intel_vgpu_active(i915)) {
> >   		DRM_INFO("GVT-g is disabled for guest\n");
> >   		goto bail;
> >   	}
> > -	if (!is_supported_device(dev_priv)) {
> > +	if (!is_supported_device(i915)) {
> >   		DRM_INFO("Unsupported device. GVT-g is disabled\n");
> >   		goto bail;
> >   	}
> > @@ -83,7 +83,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_gvt_init - initialize GVT components
> > - * @dev_priv: drm i915 private data
> > + * @i915: drm i915 private data
> >    *
> >    * This function is called at the initialization stage to create a GVT device.
> >    *
> > @@ -91,7 +91,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> >    * Zero on success, negative error code if failed.
> >    *
> >    */
> > -int intel_gvt_init(struct drm_i915_private *dev_priv)
> > +int intel_gvt_init(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > @@ -103,12 +103,12 @@ int intel_gvt_init(struct drm_i915_private *dev_priv)
> >   		return 0;
> >   	}
> > -	if (USES_GUC_SUBMISSION(dev_priv)) {
> > +	if (USES_GUC_SUBMISSION(i915)) {
> >   		DRM_ERROR("i915 GVT-g loading failed due to Graphics virtualization is not yet supported with GuC submission\n");
> >   		return -EIO;
> >   	}
> > -	ret = intel_gvt_init_device(dev_priv);
> > +	ret = intel_gvt_init_device(i915);
> >   	if (ret) {
> >   		DRM_DEBUG_DRIVER("Fail to init GVT device\n");
> >   		goto bail;
> > @@ -123,15 +123,15 @@ int intel_gvt_init(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_gvt_cleanup - cleanup GVT components when i915 driver is unloading
> > - * @dev_priv: drm i915 private *
> > + * @i915: drm i915 private *
> >    *
> >    * This function is called at the i915 driver unloading stage, to shutdown
> >    * GVT components and release the related resources.
> >    */
> > -void intel_gvt_cleanup(struct drm_i915_private *dev_priv)
> > +void intel_gvt_cleanup(struct drm_i915_private *i915)
> >   {
> > -	if (!intel_gvt_active(dev_priv))
> > +	if (!intel_gvt_active(i915))
> >   		return;
> > -	intel_gvt_clean_device(dev_priv);
> > +	intel_gvt_clean_device(i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_gvt.h b/drivers/gpu/drm/i915/intel_gvt.h
> > index 61b246470282..6f81e26cb9ad 100644
> > --- a/drivers/gpu/drm/i915/intel_gvt.h
> > +++ b/drivers/gpu/drm/i915/intel_gvt.h
> > @@ -27,22 +27,22 @@
> >   struct intel_gvt;
> >   #ifdef CONFIG_DRM_I915_GVT
> > -int intel_gvt_init(struct drm_i915_private *dev_priv);
> > -void intel_gvt_cleanup(struct drm_i915_private *dev_priv);
> > -int intel_gvt_init_device(struct drm_i915_private *dev_priv);
> > -void intel_gvt_clean_device(struct drm_i915_private *dev_priv);
> > +int intel_gvt_init(struct drm_i915_private *i915);
> > +void intel_gvt_cleanup(struct drm_i915_private *i915);
> > +int intel_gvt_init_device(struct drm_i915_private *i915);
> > +void intel_gvt_clean_device(struct drm_i915_private *i915);
> >   int intel_gvt_init_host(void);
> > -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv);
> > +void intel_gvt_sanitize_options(struct drm_i915_private *i915);
> >   #else
> > -static inline int intel_gvt_init(struct drm_i915_private *dev_priv)
> > +static inline int intel_gvt_init(struct drm_i915_private *i915)
> >   {
> >   	return 0;
> >   }
> > -static inline void intel_gvt_cleanup(struct drm_i915_private *dev_priv)
> > +static inline void intel_gvt_cleanup(struct drm_i915_private *i915)
> >   {
> >   }
> > -static inline void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv)
> > +static inline void intel_gvt_sanitize_options(struct drm_i915_private *i915)
> >   {
> >   }
> >   #endif
> > diff --git a/drivers/gpu/drm/i915/intel_hdcp.c b/drivers/gpu/drm/i915/intel_hdcp.c
> > index bc3a94d491c4..7df5f19b52d5 100644
> > --- a/drivers/gpu/drm/i915/intel_hdcp.c
> > +++ b/drivers/gpu/drm/i915/intel_hdcp.c
> > @@ -81,7 +81,7 @@ bool intel_hdcp_capable(struct intel_connector *connector)
> >   /* Is HDCP2.2 capable on Platform and Sink */
> >   bool intel_hdcp2_capable(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	bool capable = false;
> > @@ -91,12 +91,12 @@ bool intel_hdcp2_capable(struct intel_connector *connector)
> >   		return false;
> >   	/* MEI interface is solid */
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	if (!dev_priv->hdcp_comp_added ||  !dev_priv->hdcp_master) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	if (!i915->hdcp_comp_added ||  !i915->hdcp_master) {
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return false;
> >   	}
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	/* Sink's capability for HDCP2.2 */
> >   	hdcp->shim->hdcp_2_2_capable(intel_dig_port, &capable);
> > @@ -106,7 +106,7 @@ bool intel_hdcp2_capable(struct intel_connector *connector)
> >   static inline bool intel_hdcp_in_use(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum port port = connector->encoder->port;
> >   	u32 reg;
> > @@ -116,7 +116,7 @@ static inline bool intel_hdcp_in_use(struct intel_connector *connector)
> >   static inline bool intel_hdcp2_in_use(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum port port = connector->encoder->port;
> >   	u32 reg;
> > @@ -145,9 +145,9 @@ static int intel_hdcp_poll_ksv_fifo(struct intel_digital_port *intel_dig_port,
> >   	return 0;
> >   }
> > -static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> > +static bool hdcp_key_loadable(struct drm_i915_private *i915)
> >   {
> > -	struct i915_power_domains *power_domains = &dev_priv->power_domains;
> > +	struct i915_power_domains *power_domains = &i915->power_domains;
> >   	struct i915_power_well *power_well;
> >   	enum i915_power_well_id id;
> >   	bool enabled = false;
> > @@ -156,7 +156,7 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> >   	 * On HSW and BDW, Display HW loads the Key as soon as Display resumes.
> >   	 * On all BXT+, SW can load the keys only when the PW#1 is turned on.
> >   	 */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		id = HSW_DISP_PW_GLOBAL;
> >   	else
> >   		id = SKL_DISP_PW_1;
> > @@ -164,9 +164,9 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> >   	mutex_lock(&power_domains->lock);
> >   	/* PG1 (power well #1) needs to be enabled */
> > -	for_each_power_well(dev_priv, power_well) {
> > +	for_each_power_well(i915, power_well) {
> >   		if (power_well->desc->id == id) {
> > -			enabled = power_well->desc->ops->is_enabled(dev_priv,
> > +			enabled = power_well->desc->ops->is_enabled(i915,
> >   								    power_well);
> >   			break;
> >   		}
> > @@ -182,14 +182,14 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv)
> >   	return enabled;
> >   }
> > -static void intel_hdcp_clear_keys(struct drm_i915_private *dev_priv)
> > +static void intel_hdcp_clear_keys(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(HDCP_KEY_CONF, HDCP_CLEAR_KEYS_TRIGGER);
> >   	I915_WRITE(HDCP_KEY_STATUS, HDCP_KEY_LOAD_DONE | HDCP_KEY_LOAD_STATUS |
> >   		   HDCP_FUSE_IN_PROGRESS | HDCP_FUSE_ERROR | HDCP_FUSE_DONE);
> >   }
> > -static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> > +static int intel_hdcp_load_keys(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> >   	u32 val;
> > @@ -202,7 +202,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> >   	 * On HSW and BDW HW loads the HDCP1.4 Key when Display comes
> >   	 * out of reset. So if Key is not already loaded, its an error state.
> >   	 */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		if (!(I915_READ(HDCP_KEY_STATUS) & HDCP_KEY_LOAD_DONE))
> >   			return -ENXIO;
> > @@ -213,8 +213,8 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> >   	 * platforms except BXT and GLK, differ in the key load trigger process
> >   	 * from other platforms. So GEN9_BC uses the GT Driver Mailbox i/f.
> >   	 */
> > -	if (IS_GEN9_BC(dev_priv)) {
> > -		ret = sandybridge_pcode_write(dev_priv,
> > +	if (IS_GEN9_BC(i915)) {
> > +		ret = sandybridge_pcode_write(i915,
> >   					      SKL_PCODE_LOAD_HDCP_KEYS, 1);
> >   		if (ret) {
> >   			DRM_ERROR("Failed to initiate HDCP key load (%d)\n",
> > @@ -226,7 +226,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> >   	}
> >   	/* Wait for the keys to load (500us) */
> > -	ret = __intel_wait_for_register(&dev_priv->uncore, HDCP_KEY_STATUS,
> > +	ret = __intel_wait_for_register(&i915->uncore, HDCP_KEY_STATUS,
> >   					HDCP_KEY_LOAD_DONE, HDCP_KEY_LOAD_DONE,
> >   					10, 1, &val);
> >   	if (ret)
> > @@ -241,10 +241,10 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv)
> >   }
> >   /* Returns updated SHA-1 index */
> > -static int intel_write_sha_text(struct drm_i915_private *dev_priv, u32 sha_text)
> > +static int intel_write_sha_text(struct drm_i915_private *i915, u32 sha_text)
> >   {
> >   	I915_WRITE(HDCP_SHA_TEXT, sha_text);
> > -	if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL,
> > +	if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL,
> >   				    HDCP_SHA1_READY, HDCP_SHA1_READY, 1)) {
> >   		DRM_ERROR("Timed out waiting for SHA1 ready\n");
> >   		return -ETIMEDOUT;
> > @@ -279,11 +279,11 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   				const struct intel_hdcp_shim *shim,
> >   				u8 *ksv_fifo, u8 num_downstream, u8 *bstatus)
> >   {
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	u32 vprime, sha_text, sha_leftovers, rep_ctl;
> >   	int ret, i, j, sha_idx;
> > -	dev_priv = intel_dig_port->base.base.dev->dev_private;
> > +	i915 = intel_dig_port->base.base.dev->dev_private;
> >   	/* Process V' values from the receiver */
> >   	for (i = 0; i < DRM_HDCP_V_PRIME_NUM_PARTS; i++) {
> > @@ -317,7 +317,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		for (j = 0; j < sha_empty; j++)
> >   			sha_text |= ksv[j] << ((sizeof(sha_text) - j - 1) * 8);
> > -		ret = intel_write_sha_text(dev_priv, sha_text);
> > +		ret = intel_write_sha_text(i915, sha_text);
> >   		if (ret < 0)
> >   			return ret;
> > @@ -340,7 +340,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		if (sizeof(sha_text) > sha_leftovers)
> >   			continue;
> > -		ret = intel_write_sha_text(dev_priv, sha_text);
> > +		ret = intel_write_sha_text(i915, sha_text);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_leftovers = 0;
> > @@ -357,7 +357,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   	if (sha_leftovers == 0) {
> >   		/* Write 16 bits of text, 16 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16);
> > -		ret = intel_write_sha_text(dev_priv,
> > +		ret = intel_write_sha_text(i915,
> >   					   bstatus[0] << 8 | bstatus[1]);
> >   		if (ret < 0)
> >   			return ret;
> > @@ -365,14 +365,14 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		/* Write 32 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 16 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> > @@ -383,21 +383,21 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		sha_text |= bstatus[0] << 16 | bstatus[1] << 8;
> >   		/* Only 24-bits of data, must be in the LSB */
> >   		sha_text = (sha_text & 0xffffff00) >> 8;
> > -		ret = intel_write_sha_text(dev_priv, sha_text);
> > +		ret = intel_write_sha_text(i915, sha_text);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 32 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 24 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> > @@ -406,7 +406,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		/* Write 32 bits of text */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> >   		sha_text |= bstatus[0] << 24 | bstatus[1] << 16;
> > -		ret = intel_write_sha_text(dev_priv, sha_text);
> > +		ret = intel_write_sha_text(i915, sha_text);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> > @@ -414,7 +414,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		/* Write 64 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> >   		for (i = 0; i < 2; i++) {
> > -			ret = intel_write_sha_text(dev_priv, 0);
> > +			ret = intel_write_sha_text(i915, 0);
> >   			if (ret < 0)
> >   				return ret;
> >   			sha_idx += sizeof(sha_text);
> > @@ -423,28 +423,28 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   		/* Write 32 bits of text */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> >   		sha_text |= bstatus[0] << 24;
> > -		ret = intel_write_sha_text(dev_priv, sha_text);
> > +		ret = intel_write_sha_text(i915, sha_text);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 8 bits of text, 24 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8);
> > -		ret = intel_write_sha_text(dev_priv, bstatus[1]);
> > +		ret = intel_write_sha_text(i915, bstatus[1]);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 32 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> >   		/* Write 8 bits of M0 */
> >   		I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_24);
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> > @@ -457,7 +457,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   	I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32);
> >   	/* Fill up to 64-4 bytes with zeros (leave the last write for length) */
> >   	while ((sha_idx % 64) < (64 - sizeof(sha_text))) {
> > -		ret = intel_write_sha_text(dev_priv, 0);
> > +		ret = intel_write_sha_text(i915, 0);
> >   		if (ret < 0)
> >   			return ret;
> >   		sha_idx += sizeof(sha_text);
> > @@ -469,13 +469,13 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port,
> >   	 *  - 10 bytes for BINFO/BSTATUS(2), M0(8)
> >   	 */
> >   	sha_text = (num_downstream * 5 + 10) * 8;
> > -	ret = intel_write_sha_text(dev_priv, sha_text);
> > +	ret = intel_write_sha_text(i915, sha_text);
> >   	if (ret < 0)
> >   		return ret;
> >   	/* Tell the HW we're done with the hash and wait for it to ACK */
> >   	I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_COMPLETE_HASH);
> > -	if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL,
> > +	if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL,
> >   				    HDCP_SHA1_COMPLETE,
> >   				    HDCP_SHA1_COMPLETE, 1)) {
> >   		DRM_ERROR("Timed out waiting for SHA1 complete\n");
> > @@ -571,7 +571,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	struct drm_device *dev = connector->base.dev;
> >   	const struct intel_hdcp_shim *shim = hdcp->shim;
> > -	struct drm_i915_private *dev_priv;
> > +	struct drm_i915_private *i915;
> >   	enum port port;
> >   	unsigned long r0_prime_gen_start;
> >   	int ret, i, tries = 2;
> > @@ -589,7 +589,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> >   	} ri;
> >   	bool repeater_present, hdcp_capable;
> > -	dev_priv = intel_dig_port->base.base.dev->dev_private;
> > +	i915 = intel_dig_port->base.base.dev->dev_private;
> >   	port = intel_dig_port->base.port;
> > @@ -615,7 +615,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> >   	I915_WRITE(PORT_HDCP_CONF(port), HDCP_CONF_CAPTURE_AN);
> >   	/* Wait for An to be acquired */
> > -	if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port),
> > +	if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port),
> >   				    HDCP_STATUS_AN_READY,
> >   				    HDCP_STATUS_AN_READY, 1)) {
> >   		DRM_ERROR("Timed out waiting for An\n");
> > @@ -701,7 +701,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> >   	}
> >   	/* Wait for encryption confirmation */
> > -	if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port),
> > +	if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port),
> >   				    HDCP_STATUS_ENC, HDCP_STATUS_ENC,
> >   				    ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) {
> >   		DRM_ERROR("Timed out waiting for encryption\n");
> > @@ -723,7 +723,7 @@ static int intel_hdcp_auth(struct intel_connector *connector)
> >   static int _intel_hdcp_disable(struct intel_connector *connector)
> >   {
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> > -	struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> > +	struct drm_i915_private *i915 = connector->base.dev->dev_private;
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> >   	enum port port = intel_dig_port->base.port;
> >   	int ret;
> > @@ -733,7 +733,7 @@ static int _intel_hdcp_disable(struct intel_connector *connector)
> >   	hdcp->hdcp_encrypted = false;
> >   	I915_WRITE(PORT_HDCP_CONF(port), 0);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    PORT_HDCP_STATUS(port), ~0, 0,
> >   				    ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) {
> >   		DRM_ERROR("Failed to disable HDCP, timeout clearing status\n");
> > @@ -753,22 +753,22 @@ static int _intel_hdcp_disable(struct intel_connector *connector)
> >   static int _intel_hdcp_enable(struct intel_connector *connector)
> >   {
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> > -	struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> > +	struct drm_i915_private *i915 = connector->base.dev->dev_private;
> >   	int i, ret, tries = 3;
> >   	DRM_DEBUG_KMS("[%s:%d] HDCP is being enabled...\n",
> >   		      connector->base.name, connector->base.base.id);
> > -	if (!hdcp_key_loadable(dev_priv)) {
> > +	if (!hdcp_key_loadable(i915)) {
> >   		DRM_ERROR("HDCP key Load is not possible\n");
> >   		return -ENXIO;
> >   	}
> >   	for (i = 0; i < KEY_LOAD_TRIES; i++) {
> > -		ret = intel_hdcp_load_keys(dev_priv);
> > +		ret = intel_hdcp_load_keys(i915);
> >   		if (!ret)
> >   			break;
> > -		intel_hdcp_clear_keys(dev_priv);
> > +		intel_hdcp_clear_keys(i915);
> >   	}
> >   	if (ret) {
> >   		DRM_ERROR("Could not load HDCP keys, (%d)\n", ret);
> > @@ -803,7 +803,7 @@ struct intel_connector *intel_hdcp_to_connector(struct intel_hdcp *hdcp)
> >   static int intel_hdcp_check_link(struct intel_connector *connector)
> >   {
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> > -	struct drm_i915_private *dev_priv = connector->base.dev->dev_private;
> > +	struct drm_i915_private *i915 = connector->base.dev->dev_private;
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> >   	enum port port = intel_dig_port->base.port;
> >   	int ret = 0;
> > @@ -884,10 +884,10 @@ static void intel_hdcp_prop_work(struct work_struct *work)
> >   	drm_modeset_unlock(&dev->mode_config.connection_mutex);
> >   }
> > -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port)
> > +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port)
> >   {
> >   	/* PORT E doesn't have HDCP, and PORT F is disabled */
> > -	return INTEL_GEN(dev_priv) >= 9 && port < PORT_E;
> > +	return INTEL_GEN(i915) >= 9 && port < PORT_E;
> >   }
> >   static int
> > @@ -895,22 +895,22 @@ hdcp2_prepare_ake_init(struct intel_connector *connector,
> >   		       struct hdcp2_ake_init *ake_data)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->initiate_hdcp2_session(comp->mei_dev, data, ake_data);
> >   	if (ret)
> >   		DRM_DEBUG_KMS("Prepare_ake_init failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -923,15 +923,15 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector,
> >   				size_t *msg_sz)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> > @@ -940,7 +940,7 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector,
> >   							 ek_pub_km, msg_sz);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Verify rx_cert failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -949,22 +949,22 @@ static int hdcp2_verify_hprime(struct intel_connector *connector,
> >   			       struct hdcp2_ake_send_hprime *rx_hprime)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->verify_hprime(comp->mei_dev, data, rx_hprime);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Verify hprime failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -974,22 +974,22 @@ hdcp2_store_pairing_info(struct intel_connector *connector,
> >   			 struct hdcp2_ake_send_pairing_info *pairing_info)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->store_pairing_info(comp->mei_dev, data, pairing_info);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Store pairing info failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -999,22 +999,22 @@ hdcp2_prepare_lc_init(struct intel_connector *connector,
> >   		      struct hdcp2_lc_init *lc_init)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->initiate_locality_check(comp->mei_dev, data, lc_init);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Prepare lc_init failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1024,22 +1024,22 @@ hdcp2_verify_lprime(struct intel_connector *connector,
> >   		    struct hdcp2_lc_send_lprime *rx_lprime)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->verify_lprime(comp->mei_dev, data, rx_lprime);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Verify L_Prime failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1048,22 +1048,22 @@ static int hdcp2_prepare_skey(struct intel_connector *connector,
> >   			      struct hdcp2_ske_send_eks *ske_data)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->get_session_key(comp->mei_dev, data, ske_data);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Get session key failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1075,15 +1075,15 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector,
> >   				      struct hdcp2_rep_send_ack *rep_send_ack)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> > @@ -1092,7 +1092,7 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector,
> >   							 rep_send_ack);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Verify rep topology failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1102,22 +1102,22 @@ hdcp2_verify_mprime(struct intel_connector *connector,
> >   		    struct hdcp2_rep_stream_ready *stream_ready)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->verify_mprime(comp->mei_dev, data, stream_ready);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Verify mprime failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1125,43 +1125,43 @@ hdcp2_verify_mprime(struct intel_connector *connector,
> >   static int hdcp2_authenticate_port(struct intel_connector *connector)
> >   {
> >   	struct hdcp_port_data *data = &connector->hdcp.port_data;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->enable_hdcp_authentication(comp->mei_dev, data);
> >   	if (ret < 0)
> >   		DRM_DEBUG_KMS("Enable hdcp auth failed. %d\n", ret);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> >   static int hdcp2_close_mei_session(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct i915_hdcp_comp_master *comp;
> >   	int ret;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	comp = dev_priv->hdcp_master;
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	comp = i915->hdcp_master;
> >   	if (!comp || !comp->ops) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return -EINVAL;
> >   	}
> >   	ret = comp->ops->close_hdcp_session(comp->mei_dev,
> >   					     &connector->hdcp.port_data);
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return ret;
> >   }
> > @@ -1488,7 +1488,7 @@ static int hdcp2_authenticate_sink(struct intel_connector *connector)
> >   static int hdcp2_enable_encryption(struct intel_connector *connector)
> >   {
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	enum port port = connector->encoder->port;
> >   	int ret;
> > @@ -1511,7 +1511,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector)
> >   			   CTL_LINK_ENCRYPTION_REQ);
> >   	}
> > -	ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port),
> > +	ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port),
> >   				      LINK_ENCRYPTION_STATUS,
> >   				      LINK_ENCRYPTION_STATUS,
> >   				      ENCRYPT_STATUS_CHANGE_TIMEOUT_MS);
> > @@ -1522,7 +1522,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector)
> >   static int hdcp2_disable_encryption(struct intel_connector *connector)
> >   {
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	enum port port = connector->encoder->port;
> >   	int ret;
> > @@ -1532,7 +1532,7 @@ static int hdcp2_disable_encryption(struct intel_connector *connector)
> >   	I915_WRITE(HDCP2_CTL_DDI(port),
> >   		   I915_READ(HDCP2_CTL_DDI(port)) & ~CTL_LINK_ENCRYPTION_REQ);
> > -	ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port),
> > +	ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port),
> >   				      LINK_ENCRYPTION_STATUS, 0x0,
> >   				      ENCRYPT_STATUS_CHANGE_TIMEOUT_MS);
> >   	if (ret == -ETIMEDOUT)
> > @@ -1628,7 +1628,7 @@ static int _intel_hdcp2_disable(struct intel_connector *connector)
> >   static int intel_hdcp2_check_link(struct intel_connector *connector)
> >   {
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	enum port port = connector->encoder->port;
> >   	int ret = 0;
> > @@ -1721,13 +1721,13 @@ static void intel_hdcp_check_work(struct work_struct *work)
> >   static int i915_hdcp_component_bind(struct device *i915_kdev,
> >   				    struct device *mei_kdev, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
> >   	DRM_DEBUG("I915 HDCP comp bind\n");
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	dev_priv->hdcp_master = (struct i915_hdcp_comp_master *)data;
> > -	dev_priv->hdcp_master->mei_dev = mei_kdev;
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	i915->hdcp_master = (struct i915_hdcp_comp_master *)data;
> > +	i915->hdcp_master->mei_dev = mei_kdev;
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   	return 0;
> >   }
> > @@ -1735,12 +1735,12 @@ static int i915_hdcp_component_bind(struct device *i915_kdev,
> >   static void i915_hdcp_component_unbind(struct device *i915_kdev,
> >   				       struct device *mei_kdev, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev);
> > +	struct drm_i915_private *i915 = kdev_to_i915(i915_kdev);
> >   	DRM_DEBUG("I915 HDCP comp unbind\n");
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	dev_priv->hdcp_master = NULL;
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	i915->hdcp_master = NULL;
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> >   }
> >   static const struct component_ops i915_hdcp_component_ops = {
> > @@ -1773,34 +1773,34 @@ static inline int initialize_hdcp_port_data(struct intel_connector *connector)
> >   	return 0;
> >   }
> > -static bool is_hdcp2_supported(struct drm_i915_private *dev_priv)
> > +static bool is_hdcp2_supported(struct drm_i915_private *i915)
> >   {
> >   	if (!IS_ENABLED(CONFIG_INTEL_MEI_HDCP))
> >   		return false;
> > -	return (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv) ||
> > -		IS_KABYLAKE(dev_priv));
> > +	return (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915) ||
> > +		IS_KABYLAKE(i915));
> >   }
> > -void intel_hdcp_component_init(struct drm_i915_private *dev_priv)
> > +void intel_hdcp_component_init(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	if (!is_hdcp2_supported(dev_priv))
> > +	if (!is_hdcp2_supported(i915))
> >   		return;
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	WARN_ON(dev_priv->hdcp_comp_added);
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	WARN_ON(i915->hdcp_comp_added);
> > -	dev_priv->hdcp_comp_added = true;
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > -	ret = component_add_typed(dev_priv->drm.dev, &i915_hdcp_component_ops,
> > +	i915->hdcp_comp_added = true;
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> > +	ret = component_add_typed(i915->drm.dev, &i915_hdcp_component_ops,
> >   				  I915_COMPONENT_HDCP);
> >   	if (ret < 0) {
> >   		DRM_DEBUG_KMS("Failed at component add(%d)\n", ret);
> > -		mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -		dev_priv->hdcp_comp_added = false;
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +		mutex_lock(&i915->hdcp_comp_mutex);
> > +		i915->hdcp_comp_added = false;
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return;
> >   	}
> >   }
> > @@ -1822,7 +1822,7 @@ static void intel_hdcp2_init(struct intel_connector *connector)
> >   int intel_hdcp_init(struct intel_connector *connector,
> >   		    const struct intel_hdcp_shim *shim)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_hdcp *hdcp = &connector->hdcp;
> >   	int ret;
> > @@ -1838,7 +1838,7 @@ int intel_hdcp_init(struct intel_connector *connector,
> >   	INIT_DELAYED_WORK(&hdcp->check_work, intel_hdcp_check_work);
> >   	INIT_WORK(&hdcp->prop_work, intel_hdcp_prop_work);
> > -	if (is_hdcp2_supported(dev_priv))
> > +	if (is_hdcp2_supported(i915))
> >   		intel_hdcp2_init(connector);
> >   	init_waitqueue_head(&hdcp->cp_irq_queue);
> > @@ -1905,18 +1905,18 @@ int intel_hdcp_disable(struct intel_connector *connector)
> >   	return ret;
> >   }
> > -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv)
> > +void intel_hdcp_component_fini(struct drm_i915_private *i915)
> >   {
> > -	mutex_lock(&dev_priv->hdcp_comp_mutex);
> > -	if (!dev_priv->hdcp_comp_added) {
> > -		mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	mutex_lock(&i915->hdcp_comp_mutex);
> > +	if (!i915->hdcp_comp_added) {
> > +		mutex_unlock(&i915->hdcp_comp_mutex);
> >   		return;
> >   	}
> > -	dev_priv->hdcp_comp_added = false;
> > -	mutex_unlock(&dev_priv->hdcp_comp_mutex);
> > +	i915->hdcp_comp_added = false;
> > +	mutex_unlock(&i915->hdcp_comp_mutex);
> > -	component_del(dev_priv->drm.dev, &i915_hdcp_component_ops);
> > +	component_del(i915->drm.dev, &i915_hdcp_component_ops);
> >   }
> >   void intel_hdcp_cleanup(struct intel_connector *connector)
> > diff --git a/drivers/gpu/drm/i915/intel_hdcp.h b/drivers/gpu/drm/i915/intel_hdcp.h
> > index be8da85c866a..24079dda1ac4 100644
> > --- a/drivers/gpu/drm/i915/intel_hdcp.h
> > +++ b/drivers/gpu/drm/i915/intel_hdcp.h
> > @@ -23,11 +23,11 @@ int intel_hdcp_init(struct intel_connector *connector,
> >   		    const struct intel_hdcp_shim *hdcp_shim);
> >   int intel_hdcp_enable(struct intel_connector *connector);
> >   int intel_hdcp_disable(struct intel_connector *connector);
> > -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port);
> > +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port);
> >   bool intel_hdcp_capable(struct intel_connector *connector);
> >   bool intel_hdcp2_capable(struct intel_connector *connector);
> > -void intel_hdcp_component_init(struct drm_i915_private *dev_priv);
> > -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv);
> > +void intel_hdcp_component_init(struct drm_i915_private *i915);
> > +void intel_hdcp_component_fini(struct drm_i915_private *i915);
> >   void intel_hdcp_cleanup(struct intel_connector *connector);
> >   void intel_hdcp_handle_cp_irq(struct intel_connector *connector);
> > diff --git a/drivers/gpu/drm/i915/intel_hdmi.c b/drivers/gpu/drm/i915/intel_hdmi.c
> > index 187a2b828b97..c5be41ba36da 100644
> > --- a/drivers/gpu/drm/i915/intel_hdmi.c
> > +++ b/drivers/gpu/drm/i915/intel_hdmi.c
> > @@ -67,17 +67,17 @@ static void
> >   assert_hdmi_port_disabled(struct intel_hdmi *intel_hdmi)
> >   {
> >   	struct drm_device *dev = intel_hdmi_to_dev(intel_hdmi);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 enabled_bits;
> > -	enabled_bits = HAS_DDI(dev_priv) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE;
> > +	enabled_bits = HAS_DDI(i915) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE;
> >   	WARN(I915_READ(intel_hdmi->hdmi_reg) & enabled_bits,
> >   	     "HDMI port enabled, expecting disabled\n");
> >   }
> >   static void
> > -assert_hdmi_transcoder_func_disabled(struct drm_i915_private *dev_priv,
> > +assert_hdmi_transcoder_func_disabled(struct drm_i915_private *i915,
> >   				     enum transcoder cpu_transcoder)
> >   {
> >   	WARN(I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder)) &
> > @@ -163,7 +163,7 @@ static u32 hsw_infoframe_enable(unsigned int type)
> >   }
> >   static i915_reg_t
> > -hsw_dip_data_reg(struct drm_i915_private *dev_priv,
> > +hsw_dip_data_reg(struct drm_i915_private *i915,
> >   		 enum transcoder cpu_transcoder,
> >   		 unsigned int type,
> >   		 int i)
> > @@ -207,7 +207,7 @@ static void g4x_write_infoframe(struct intel_encoder *encoder,
> >   				const void *frame, ssize_t len)
> >   {
> >   	const u32 *data = frame;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val = I915_READ(VIDEO_DIP_CTL);
> >   	int i;
> > @@ -241,7 +241,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder,
> >   			       unsigned int type,
> >   			       void *frame, ssize_t len)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val, *data = frame;
> >   	int i;
> > @@ -259,7 +259,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder,
> >   static u32 g4x_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val = I915_READ(VIDEO_DIP_CTL);
> >   	if ((val & VIDEO_DIP_ENABLE) == 0)
> > @@ -278,7 +278,7 @@ static void ibx_write_infoframe(struct intel_encoder *encoder,
> >   				const void *frame, ssize_t len)
> >   {
> >   	const u32 *data = frame;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> >   	u32 val = I915_READ(reg);
> > @@ -314,7 +314,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder,
> >   			       unsigned int type,
> >   			       void *frame, ssize_t len)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	u32 val, *data = frame;
> >   	int i;
> > @@ -333,7 +333,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder,
> >   static u32 ibx_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> >   	i915_reg_t reg = TVIDEO_DIP_CTL(pipe);
> >   	u32 val = I915_READ(reg);
> > @@ -355,7 +355,7 @@ static void cpt_write_infoframe(struct intel_encoder *encoder,
> >   				const void *frame, ssize_t len)
> >   {
> >   	const u32 *data = frame;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> >   	u32 val = I915_READ(reg);
> > @@ -394,7 +394,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder,
> >   			       unsigned int type,
> >   			       void *frame, ssize_t len)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	u32 val, *data = frame;
> >   	int i;
> > @@ -413,7 +413,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder,
> >   static u32 cpt_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> >   	u32 val = I915_READ(TVIDEO_DIP_CTL(pipe));
> > @@ -431,7 +431,7 @@ static void vlv_write_infoframe(struct intel_encoder *encoder,
> >   				const void *frame, ssize_t len)
> >   {
> >   	const u32 *data = frame;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe);
> >   	u32 val = I915_READ(reg);
> > @@ -467,7 +467,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder,
> >   			       unsigned int type,
> >   			       void *frame, ssize_t len)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	u32 val, *data = frame;
> >   	int i;
> > @@ -486,7 +486,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder,
> >   static u32 vlv_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe;
> >   	u32 val = I915_READ(VLV_TVIDEO_DIP_CTL(pipe));
> > @@ -507,7 +507,7 @@ static void hsw_write_infoframe(struct intel_encoder *encoder,
> >   				const void *frame, ssize_t len)
> >   {
> >   	const u32 *data = frame;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	i915_reg_t ctl_reg = HSW_TVIDEO_DIP_CTL(cpu_transcoder);
> >   	int data_size;
> > @@ -520,13 +520,13 @@ static void hsw_write_infoframe(struct intel_encoder *encoder,
> >   	I915_WRITE(ctl_reg, val);
> >   	for (i = 0; i < len; i += 4) {
> > -		I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> > +		I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder,
> >   					    type, i >> 2), *data);
> >   		data++;
> >   	}
> >   	/* Write every possible data byte to force correct ECC calculation. */
> >   	for (; i < data_size; i += 4)
> > -		I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> > +		I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder,
> >   					    type, i >> 2), 0);
> >   	val |= hsw_infoframe_enable(type);
> > @@ -539,7 +539,7 @@ static void hsw_read_infoframe(struct intel_encoder *encoder,
> >   			       unsigned int type,
> >   			       void *frame, ssize_t len)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 val, *data = frame;
> >   	int i;
> > @@ -547,14 +547,14 @@ static void hsw_read_infoframe(struct intel_encoder *encoder,
> >   	val = I915_READ(HSW_TVIDEO_DIP_CTL(cpu_transcoder));
> >   	for (i = 0; i < len; i += 4)
> > -		*data++ = I915_READ(hsw_dip_data_reg(dev_priv, cpu_transcoder,
> > +		*data++ = I915_READ(hsw_dip_data_reg(i915, cpu_transcoder,
> >   						     type, i >> 2));
> >   }
> >   static u32 hsw_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val = I915_READ(HSW_TVIDEO_DIP_CTL(pipe_config->cpu_transcoder));
> >   	u32 mask;
> > @@ -562,7 +562,7 @@ static u32 hsw_infoframes_enabled(struct intel_encoder *encoder,
> >   		VIDEO_DIP_ENABLE_GCP_HSW | VIDEO_DIP_ENABLE_VS_HSW |
> >   		VIDEO_DIP_ENABLE_GMP_HSW | VIDEO_DIP_ENABLE_SPD_HSW);
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		mask |= VIDEO_DIP_ENABLE_DRM_GLK;
> >   	return val & mask;
> > @@ -593,7 +593,7 @@ u32 intel_hdmi_infoframe_enable(unsigned int type)
> >   u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base);
> >   	u32 val, ret = 0;
> >   	int i;
> > @@ -604,7 +604,7 @@ u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder,
> >   	for (i = 0; i < ARRAY_SIZE(infoframe_type_to_idx); i++) {
> >   		unsigned int type = infoframe_type_to_idx[i];
> > -		if (HAS_DDI(dev_priv)) {
> > +		if (HAS_DDI(i915)) {
> >   			if (val & hsw_infoframe_enable(type))
> >   				ret |= BIT(i);
> >   		} else {
> > @@ -804,10 +804,10 @@ intel_hdmi_compute_drm_infoframe(struct intel_encoder *encoder,
> >   				 struct drm_connector_state *conn_state)
> >   {
> >   	struct hdmi_drm_infoframe *frame = &crtc_state->infoframes.drm.drm;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	int ret;
> > -	if (!(INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)))
> > +	if (!(INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)))
> >   		return true;
> >   	if (!crtc_state->has_infoframe)
> > @@ -837,7 +837,7 @@ static void g4x_set_infoframes(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> >   	struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> >   	i915_reg_t reg = VIDEO_DIP_CTL;
> > @@ -947,7 +947,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> >   					 const struct intel_crtc_state *crtc_state,
> >   					 const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	i915_reg_t reg;
> > @@ -955,11 +955,11 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> >   	     intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0)
> >   		return false;
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		reg = VLV_TVIDEO_DIP_GCP(crtc->pipe);
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > +	else if (HAS_PCH_SPLIT(i915))
> >   		reg = TVIDEO_DIP_GCP(crtc->pipe);
> >   	else
> >   		return false;
> > @@ -972,7 +972,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder,
> >   void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	i915_reg_t reg;
> > @@ -980,11 +980,11 @@ void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder,
> >   	     intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0)
> >   		return;
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> >   		reg = VLV_TVIDEO_DIP_GCP(crtc->pipe);
> > -	else if (HAS_PCH_SPLIT(dev_priv))
> > +	else if (HAS_PCH_SPLIT(i915))
> >   		reg = TVIDEO_DIP_GCP(crtc->pipe);
> >   	else
> >   		return;
> > @@ -996,9 +996,9 @@ static void intel_hdmi_compute_gcp_infoframe(struct intel_encoder *encoder,
> >   					     struct intel_crtc_state *crtc_state,
> >   					     struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > -	if (IS_G4X(dev_priv) || !crtc_state->has_infoframe)
> > +	if (IS_G4X(i915) || !crtc_state->has_infoframe)
> >   		return;
> >   	crtc_state->infoframes.enable |=
> > @@ -1019,7 +1019,7 @@ static void ibx_set_infoframes(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base);
> >   	struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> > @@ -1078,7 +1078,7 @@ static void cpt_set_infoframes(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe);
> > @@ -1127,7 +1127,7 @@ static void vlv_set_infoframes(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe);
> > @@ -1185,11 +1185,11 @@ static void hsw_set_infoframes(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	i915_reg_t reg = HSW_TVIDEO_DIP_CTL(crtc_state->cpu_transcoder);
> >   	u32 val = I915_READ(reg);
> > -	assert_hdmi_transcoder_func_disabled(dev_priv,
> > +	assert_hdmi_transcoder_func_disabled(i915,
> >   					     crtc_state->cpu_transcoder);
> >   	val &= ~(VIDEO_DIP_ENABLE_VSC_HSW | VIDEO_DIP_ENABLE_AVI_HSW |
> > @@ -1225,9 +1225,9 @@ static void hsw_set_infoframes(struct intel_encoder *encoder,
> >   void intel_dp_dual_mode_set_tmds_output(struct intel_hdmi *hdmi, bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi));
> > +	struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi));
> >   	struct i2c_adapter *adapter =
> > -		intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> > +		intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
> >   	if (hdmi->dp_dual_mode.type < DRM_DP_DUAL_MODE_TYPE2_DVI)
> >   		return;
> > @@ -1243,9 +1243,9 @@ static int intel_hdmi_hdcp_read(struct intel_digital_port *intel_dig_port,
> >   				unsigned int offset, void *buffer, size_t size)
> >   {
> >   	struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		intel_dig_port->base.base.dev->dev_private;
> > -	struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> > +	struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> >   							      hdmi->ddc_bus);
> >   	int ret;
> >   	u8 start = offset & 0xff;
> > @@ -1273,9 +1273,9 @@ static int intel_hdmi_hdcp_write(struct intel_digital_port *intel_dig_port,
> >   				 unsigned int offset, void *buffer, size_t size)
> >   {
> >   	struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		intel_dig_port->base.base.dev->dev_private;
> > -	struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> > +	struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> >   							      hdmi->ddc_bus);
> >   	int ret;
> >   	u8 *write_buf;
> > @@ -1308,9 +1308,9 @@ int intel_hdmi_hdcp_write_an_aksv(struct intel_digital_port *intel_dig_port,
> >   				  u8 *an)
> >   {
> >   	struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		intel_dig_port->base.base.dev->dev_private;
> > -	struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv,
> > +	struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915,
> >   							      hdmi->ddc_bus);
> >   	int ret;
> > @@ -1428,7 +1428,7 @@ int intel_hdmi_hdcp_read_v_prime_part(struct intel_digital_port *intel_dig_port,
> >   static int kbl_repositioning_enc_en_signal(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector);
> >   	struct drm_crtc *crtc = connector->base.state->crtc;
> >   	struct intel_crtc *intel_crtc = container_of(crtc,
> > @@ -1463,7 +1463,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> >   {
> >   	struct intel_hdmi *hdmi = &intel_dig_port->hdmi;
> >   	struct intel_connector *connector = hdmi->attached_connector;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	int ret;
> >   	if (!enable)
> > @@ -1480,7 +1480,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> >   	 * WA: To fix incorrect positioning of the window of
> >   	 * opportunity and enc_en signalling in KABYLAKE.
> >   	 */
> > -	if (IS_KABYLAKE(dev_priv) && enable)
> > +	if (IS_KABYLAKE(i915) && enable)
> >   		return kbl_repositioning_enc_en_signal(connector);
> >   	return 0;
> > @@ -1489,7 +1489,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port,
> >   static
> >   bool intel_hdmi_hdcp_check_link(struct intel_digital_port *intel_dig_port)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		intel_dig_port->base.base.dev->dev_private;
> >   	enum port port = intel_dig_port->base.port;
> >   	int ret;
> > @@ -1720,7 +1720,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> > @@ -1729,7 +1729,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> >   	intel_dp_dual_mode_set_tmds_output(intel_hdmi, true);
> >   	hdmi_val = SDVO_ENCODING_HDMI;
> > -	if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range)
> > +	if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range)
> >   		hdmi_val |= HDMI_COLOR_RANGE_16_235;
> >   	if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC)
> >   		hdmi_val |= SDVO_VSYNC_ACTIVE_HIGH;
> > @@ -1744,9 +1744,9 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> >   	if (crtc_state->has_hdmi_sink)
> >   		hdmi_val |= HDMI_MODE_SELECT_HDMI;
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		hdmi_val |= SDVO_PIPE_SEL_CPT(crtc->pipe);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915))
> >   		hdmi_val |= SDVO_PIPE_SEL_CHV(crtc->pipe);
> >   	else
> >   		hdmi_val |= SDVO_PIPE_SEL(crtc->pipe);
> > @@ -1758,19 +1758,19 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder,
> >   static bool intel_hdmi_get_hw_state(struct intel_encoder *encoder,
> >   				    enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > -	ret = intel_sdvo_port_enabled(dev_priv, intel_hdmi->hdmi_reg, pipe);
> > +	ret = intel_sdvo_port_enabled(i915, intel_hdmi->hdmi_reg, pipe);
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -1780,7 +1780,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder,
> >   {
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 tmp, flags = 0;
> >   	int dotclock;
> > @@ -1810,7 +1810,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder,
> >   	if (tmp & HDMI_AUDIO_ENABLE)
> >   		pipe_config->has_audio = true;
> > -	if (!HAS_PCH_SPLIT(dev_priv) &&
> > +	if (!HAS_PCH_SPLIT(i915) &&
> >   	    tmp & HDMI_COLOR_RANGE_16_235)
> >   		pipe_config->limited_color_range = true;
> > @@ -1858,7 +1858,7 @@ static void g4x_enable_hdmi(struct intel_encoder *encoder,
> >   			    const struct drm_connector_state *conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	u32 temp;
> > @@ -1880,7 +1880,7 @@ static void ibx_enable_hdmi(struct intel_encoder *encoder,
> >   			    const struct drm_connector_state *conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	u32 temp;
> > @@ -1930,7 +1930,7 @@ static void cpt_enable_hdmi(struct intel_encoder *encoder,
> >   			    const struct drm_connector_state *conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	enum pipe pipe = crtc->pipe;
> > @@ -1991,7 +1991,7 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> >   			       const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	struct intel_digital_port *intel_dig_port =
> >   		hdmi_to_dig_port(intel_hdmi);
> > @@ -2009,13 +2009,13 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> >   	 * to transcoder A after disabling it to allow the
> >   	 * matching DP port to be enabled on transcoder A.
> >   	 */
> > -	if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) {
> > +	if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) {
> >   		/*
> >   		 * We get CPU/PCH FIFO underruns on the other pipe when
> >   		 * doing the workaround. Sweep them under the rug.
> >   		 */
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> >   		temp &= ~SDVO_PIPE_SEL_MASK;
> >   		temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A);
> > @@ -2032,9 +2032,9 @@ static void intel_disable_hdmi(struct intel_encoder *encoder,
> >   		I915_WRITE(intel_hdmi->hdmi_reg, temp);
> >   		POSTING_READ(intel_hdmi->hdmi_reg);
> > -		intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > +		intel_wait_for_vblank_if_active(i915, PIPE_A);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> >   	}
> >   	intel_dig_port->set_infoframes(encoder,
> > @@ -2073,16 +2073,16 @@ static void pch_post_disable_hdmi(struct intel_encoder *encoder,
> >   static int intel_hdmi_source_max_tmds_clock(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	const struct ddi_vbt_port_info *info =
> > -		&dev_priv->vbt.ddi_port_info[encoder->port];
> > +		&i915->vbt.ddi_port_info[encoder->port];
> >   	int max_tmds_clock;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		max_tmds_clock = 594000;
> > -	else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv))
> > +	else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915))
> >   		max_tmds_clock = 300000;
> > -	else if (INTEL_GEN(dev_priv) >= 5)
> > +	else if (INTEL_GEN(i915) >= 5)
> >   		max_tmds_clock = 225000;
> >   	else
> >   		max_tmds_clock = 165000;
> > @@ -2123,7 +2123,7 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi,
> >   		      int clock, bool respect_downstream_limits,
> >   		      bool force_dvi)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi));
> > +	struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi));
> >   	if (clock < 25000)
> >   		return MODE_CLOCK_LOW;
> > @@ -2131,11 +2131,11 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi,
> >   		return MODE_CLOCK_HIGH;
> >   	/* BXT DPLL can't generate 223-240 MHz */
> > -	if (IS_GEN9_LP(dev_priv) && clock > 223333 && clock < 240000)
> > +	if (IS_GEN9_LP(i915) && clock > 223333 && clock < 240000)
> >   		return MODE_CLOCK_RANGE;
> >   	/* CHV DPLL can't generate 216-240 MHz */
> > -	if (IS_CHERRYVIEW(dev_priv) && clock > 216000 && clock < 240000)
> > +	if (IS_CHERRYVIEW(i915) && clock > 216000 && clock < 240000)
> >   		return MODE_CLOCK_RANGE;
> >   	return MODE_OK;
> > @@ -2147,7 +2147,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
> >   {
> >   	struct intel_hdmi *hdmi = intel_attached_hdmi(connector);
> >   	struct drm_device *dev = intel_hdmi_to_dev(hdmi);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum drm_mode_status status;
> >   	int clock;
> >   	int max_dotclk = to_i915(connector->dev)->max_dotclk_freq;
> > @@ -2176,12 +2176,12 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
> >   	if (hdmi->has_hdmi_sink && !force_dvi) {
> >   		/* if we can't do 8bpc we may still be able to do 12bpc */
> > -		if (status != MODE_OK && !HAS_GMCH(dev_priv))
> > +		if (status != MODE_OK && !HAS_GMCH(i915))
> >   			status = hdmi_port_clock_valid(hdmi, clock * 3 / 2,
> >   						       true, force_dvi);
> >   		/* if we can't do 8,12bpc we may still be able to do 10bpc */
> > -		if (status != MODE_OK && INTEL_GEN(dev_priv) >= 11)
> > +		if (status != MODE_OK && INTEL_GEN(i915) >= 11)
> >   			status = hdmi_port_clock_valid(hdmi, clock * 5 / 4,
> >   						       true, force_dvi);
> >   	}
> > @@ -2192,7 +2192,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector,
> >   static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> >   				     int bpc)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(crtc_state->base.crtc->dev);
> >   	struct drm_atomic_state *state = crtc_state->base.state;
> >   	struct drm_connector_state *connector_state;
> > @@ -2201,10 +2201,10 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> >   		&crtc_state->base.adjusted_mode;
> >   	int i;
> > -	if (HAS_GMCH(dev_priv))
> > +	if (HAS_GMCH(i915))
> >   		return false;
> > -	if (bpc == 10 && INTEL_GEN(dev_priv) < 11)
> > +	if (bpc == 10 && INTEL_GEN(i915) < 11)
> >   		return false;
> >   	if (crtc_state->pipe_bpp < bpc * 3)
> > @@ -2246,13 +2246,13 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state,
> >   	}
> >   	/* Display WA #1139: glk */
> > -	if (bpc == 12 && IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1) &&
> > +	if (bpc == 12 && IS_GLK_REVID(i915, 0, GLK_REVID_A1) &&
> >   	    adjusted_mode->htotal > 5460)
> >   		return false;
> >   	/* Display Wa_1405510057:icl */
> >   	if (crtc_state->output_format == INTEL_OUTPUT_FORMAT_YCBCR420 &&
> > -	    bpc == 10 && INTEL_GEN(dev_priv) >= 11 &&
> > +	    bpc == 10 && INTEL_GEN(i915) >= 11 &&
> >   	    (adjusted_mode->crtc_hblank_end -
> >   	     adjusted_mode->crtc_hblank_start) % 8 == 2)
> >   		return false;
> > @@ -2297,7 +2297,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
> >   			      struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> >   	struct drm_connector *connector = conn_state->connector;
> >   	struct drm_scdc *scdc = &connector->display_info.hdmi.scdc;
> > @@ -2345,7 +2345,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
> >   		}
> >   	}
> > -	if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915))
> >   		pipe_config->has_pch_encoder = true;
> >   	if (pipe_config->has_hdmi_sink) {
> > @@ -2399,8 +2399,8 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder,
> >   	pipe_config->lane_count = 4;
> > -	if (scdc->scrambling.supported && (INTEL_GEN(dev_priv) >= 10 ||
> > -					   IS_GEMINILAKE(dev_priv))) {
> > +	if (scdc->scrambling.supported && (INTEL_GEN(i915) >= 10 ||
> > +					   IS_GEMINILAKE(i915))) {
> >   		if (scdc->scrambling.low_rates)
> >   			pipe_config->hdmi_scrambling = true;
> > @@ -2453,11 +2453,11 @@ intel_hdmi_unset_edid(struct drm_connector *connector)
> >   static void
> >   intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_hdmi *hdmi = intel_attached_hdmi(connector);
> >   	enum port port = hdmi_to_dig_port(hdmi)->base.port;
> >   	struct i2c_adapter *adapter =
> > -		intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus);
> > +		intel_gmbus_get_adapter(i915, hdmi->ddc_bus);
> >   	enum drm_dp_dual_mode_type type = drm_dp_dual_mode_detect(adapter);
> >   	/*
> > @@ -2477,7 +2477,7 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> >   		 * Make sure not to set limits for that port.
> >   		 */
> >   		if (has_edid && !connector->override_edid &&
> > -		    intel_bios_is_port_dp_dual_mode(dev_priv, port)) {
> > +		    intel_bios_is_port_dp_dual_mode(i915, port)) {
> >   			DRM_DEBUG_KMS("Assuming DP dual mode adaptor presence based on VBT\n");
> >   			type = DRM_DP_DUAL_MODE_TYPE1_DVI;
> >   		} else {
> > @@ -2500,16 +2500,16 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid)
> >   static bool
> >   intel_hdmi_set_edid(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
> >   	intel_wakeref_t wakeref;
> >   	struct edid *edid;
> >   	bool connected = false;
> >   	struct i2c_adapter *i2c;
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
> > -	i2c = intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> > +	i2c = intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
> >   	edid = drm_get_edid(connector, i2c);
> > @@ -2522,7 +2522,7 @@ intel_hdmi_set_edid(struct drm_connector *connector)
> >   	intel_hdmi_dp_dual_mode_detect(connector, edid != NULL);
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
> >   	to_intel_connector(connector)->detect_edid = edid;
> >   	if (edid && edid->input & DRM_EDID_INPUT_DIGITAL) {
> > @@ -2541,7 +2541,7 @@ static enum drm_connector_status
> >   intel_hdmi_detect(struct drm_connector *connector, bool force)
> >   {
> >   	enum drm_connector_status status = connector_status_disconnected;
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
> >   	struct intel_encoder *encoder = &hdmi_to_dig_port(intel_hdmi)->base;
> >   	intel_wakeref_t wakeref;
> > @@ -2549,9 +2549,9 @@ intel_hdmi_detect(struct drm_connector *connector, bool force)
> >   	DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> >   		      connector->base.id, connector->name);
> > -	wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS);
> > +	wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS);
> > -	if (INTEL_GEN(dev_priv) >= 11 &&
> > +	if (INTEL_GEN(i915) >= 11 &&
> >   	    !intel_digital_port_connected(encoder))
> >   		goto out;
> > @@ -2561,7 +2561,7 @@ intel_hdmi_detect(struct drm_connector *connector, bool force)
> >   		status = connector_status_connected;
> >   out:
> > -	intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref);
> > +	intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref);
> >   	if (status != connector_status_connected)
> >   		cec_notifier_phys_addr_invalidate(intel_hdmi->cec_notifier);
> > @@ -2613,7 +2613,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder,
> >   				const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	vlv_phy_pre_encoder_enable(encoder, pipe_config);
> > @@ -2627,7 +2627,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder,
> >   	g4x_enable_hdmi(encoder, pipe_config, conn_state);
> > -	vlv_wait_port_ready(dev_priv, dport, 0x0);
> > +	vlv_wait_port_ready(i915, dport, 0x0);
> >   }
> >   static void vlv_hdmi_pre_pll_enable(struct intel_encoder *encoder,
> > @@ -2668,14 +2668,14 @@ static void chv_hdmi_post_disable(struct intel_encoder *encoder,
> >   				  const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	vlv_dpio_get(dev_priv);
> > +	vlv_dpio_get(i915);
> >   	/* Assert data lane reset */
> >   	chv_data_lane_soft_reset(encoder, old_crtc_state, true);
> > -	vlv_dpio_put(dev_priv);
> > +	vlv_dpio_put(i915);
> >   }
> >   static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> > @@ -2684,7 +2684,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> >   {
> >   	struct intel_digital_port *dport = enc_to_dig_port(&encoder->base);
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	chv_phy_pre_encoder_enable(encoder, pipe_config);
> > @@ -2698,7 +2698,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> >   	g4x_enable_hdmi(encoder, pipe_config, conn_state);
> > -	vlv_wait_port_ready(dev_priv, dport, 0x0);
> > +	vlv_wait_port_ready(i915, dport, 0x0);
> >   	/* Second common lane will stay alive on its own now */
> >   	chv_phy_release_cl2_override(encoder);
> > @@ -2707,10 +2707,10 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder,
> >   static struct i2c_adapter *
> >   intel_hdmi_get_i2c_adapter(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector);
> > -	return intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> > +	return intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
> >   }
> >   static void intel_hdmi_create_i2c_symlink(struct drm_connector *connector)
> > @@ -2791,7 +2791,7 @@ static const struct drm_encoder_funcs intel_hdmi_enc_funcs = {
> >   static void
> >   intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_digital_port *intel_dig_port =
> >   				hdmi_to_dig_port(intel_hdmi);
> > @@ -2810,11 +2810,11 @@ intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *c
> >   	drm_connector_attach_content_type_property(connector);
> >   	connector->state->picture_aspect_ratio = HDMI_PICTURE_ASPECT_NONE;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		drm_object_attach_property(&connector->base,
> >   			connector->dev->mode_config.hdr_output_metadata_property, 0);
> > -	if (!HAS_GMCH(dev_priv))
> > +	if (!HAS_GMCH(i915))
> >   		drm_connector_attach_max_bpc_property(connector, 8, 12);
> >   }
> > @@ -2841,12 +2841,12 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder,
> >   				       bool high_tmds_clock_ratio,
> >   				       bool scrambling)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base);
> >   	struct drm_scrambling *sink_scrambling =
> >   		&connector->display_info.hdmi.scdc.scrambling;
> >   	struct i2c_adapter *adapter =
> > -		intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus);
> > +		intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus);
> >   	if (!sink_scrambling->supported)
> >   		return true;
> > @@ -2861,7 +2861,7 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder,
> >   		drm_scdc_set_scrambling(adapter, scrambling);
> >   }
> > -static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> > +static u8 chv_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> >   {
> >   	u8 ddc_pin;
> > @@ -2883,7 +2883,7 @@ static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> >   	return ddc_pin;
> >   }
> > -static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> > +static u8 bxt_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> >   {
> >   	u8 ddc_pin;
> > @@ -2902,7 +2902,7 @@ static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> >   	return ddc_pin;
> >   }
> > -static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> > +static u8 cnp_port_to_ddc_pin(struct drm_i915_private *i915,
> >   			      enum port port)
> >   {
> >   	u8 ddc_pin;
> > @@ -2928,7 +2928,7 @@ static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> >   	return ddc_pin;
> >   }
> > -static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> > +static u8 icl_port_to_ddc_pin(struct drm_i915_private *i915, enum port port)
> >   {
> >   	u8 ddc_pin;
> > @@ -2959,7 +2959,7 @@ static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port)
> >   	return ddc_pin;
> >   }
> > -static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> > +static u8 g4x_port_to_ddc_pin(struct drm_i915_private *i915,
> >   			      enum port port)
> >   {
> >   	u8 ddc_pin;
> > @@ -2982,11 +2982,11 @@ static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv,
> >   	return ddc_pin;
> >   }
> > -static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
> > +static u8 intel_hdmi_ddc_pin(struct drm_i915_private *i915,
> >   			     enum port port)
> >   {
> >   	const struct ddi_vbt_port_info *info =
> > -		&dev_priv->vbt.ddi_port_info[port];
> > +		&i915->vbt.ddi_port_info[port];
> >   	u8 ddc_pin;
> >   	if (info->alternate_ddc_pin) {
> > @@ -2995,16 +2995,16 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
> >   		return info->alternate_ddc_pin;
> >   	}
> > -	if (HAS_PCH_ICP(dev_priv))
> > -		ddc_pin = icl_port_to_ddc_pin(dev_priv, port);
> > -	else if (HAS_PCH_CNP(dev_priv))
> > -		ddc_pin = cnp_port_to_ddc_pin(dev_priv, port);
> > -	else if (IS_GEN9_LP(dev_priv))
> > -		ddc_pin = bxt_port_to_ddc_pin(dev_priv, port);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		ddc_pin = chv_port_to_ddc_pin(dev_priv, port);
> > +	if (HAS_PCH_ICP(i915))
> > +		ddc_pin = icl_port_to_ddc_pin(i915, port);
> > +	else if (HAS_PCH_CNP(i915))
> > +		ddc_pin = cnp_port_to_ddc_pin(i915, port);
> > +	else if (IS_GEN9_LP(i915))
> > +		ddc_pin = bxt_port_to_ddc_pin(i915, port);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		ddc_pin = chv_port_to_ddc_pin(i915, port);
> >   	else
> > -		ddc_pin = g4x_port_to_ddc_pin(dev_priv, port);
> > +		ddc_pin = g4x_port_to_ddc_pin(i915, port);
> >   	DRM_DEBUG_KMS("Using DDC pin 0x%x for port %c (platform default)\n",
> >   		      ddc_pin, port_name(port));
> > @@ -3014,20 +3014,20 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv,
> >   void intel_infoframe_init(struct intel_digital_port *intel_dig_port)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(intel_dig_port->base.base.dev);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		intel_dig_port->write_infoframe = vlv_write_infoframe;
> >   		intel_dig_port->read_infoframe = vlv_read_infoframe;
> >   		intel_dig_port->set_infoframes = vlv_set_infoframes;
> >   		intel_dig_port->infoframes_enabled = vlv_infoframes_enabled;
> > -	} else if (IS_G4X(dev_priv)) {
> > +	} else if (IS_G4X(i915)) {
> >   		intel_dig_port->write_infoframe = g4x_write_infoframe;
> >   		intel_dig_port->read_infoframe = g4x_read_infoframe;
> >   		intel_dig_port->set_infoframes = g4x_set_infoframes;
> >   		intel_dig_port->infoframes_enabled = g4x_infoframes_enabled;
> > -	} else if (HAS_DDI(dev_priv)) {
> > +	} else if (HAS_DDI(i915)) {
> >   		if (intel_dig_port->lspcon.active) {
> >   			intel_dig_port->write_infoframe = lspcon_write_infoframe;
> >   			intel_dig_port->read_infoframe = lspcon_read_infoframe;
> > @@ -3039,7 +3039,7 @@ void intel_infoframe_init(struct intel_digital_port *intel_dig_port)
> >   			intel_dig_port->set_infoframes = hsw_set_infoframes;
> >   			intel_dig_port->infoframes_enabled = hsw_infoframes_enabled;
> >   		}
> > -	} else if (HAS_PCH_IBX(dev_priv)) {
> > +	} else if (HAS_PCH_IBX(i915)) {
> >   		intel_dig_port->write_infoframe = ibx_write_infoframe;
> >   		intel_dig_port->read_infoframe = ibx_read_infoframe;
> >   		intel_dig_port->set_infoframes = ibx_set_infoframes;
> > @@ -3059,7 +3059,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   	struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi;
> >   	struct intel_encoder *intel_encoder = &intel_dig_port->base;
> >   	struct drm_device *dev = intel_encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum port port = intel_encoder->port;
> >   	DRM_DEBUG_KMS("Adding HDMI connector on port %c\n",
> > @@ -3078,16 +3078,16 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   	connector->doublescan_allowed = 0;
> >   	connector->stereo_allowed = 1;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		connector->ycbcr_420_allowed = true;
> > -	intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(dev_priv, port);
> > +	intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(i915, port);
> >   	if (WARN_ON(port == PORT_A))
> >   		return;
> > -	intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port);
> > +	intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port);
> > -	if (HAS_DDI(dev_priv))
> > +	if (HAS_DDI(i915))
> >   		intel_connector->get_hw_state = intel_ddi_connector_get_hw_state;
> >   	else
> >   		intel_connector->get_hw_state = intel_connector_get_hw_state;
> > @@ -3097,7 +3097,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   	intel_connector_attach_encoder(intel_connector, intel_encoder);
> >   	intel_hdmi->attached_connector = intel_connector;
> > -	if (is_hdcp_supported(dev_priv, port)) {
> > +	if (is_hdcp_supported(i915, port)) {
> >   		int ret = intel_hdcp_init(intel_connector,
> >   					  &intel_hdmi_hdcp_shim);
> >   		if (ret)
> > @@ -3108,7 +3108,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   	 * 0xd.  Failure to do so will result in spurious interrupts being
> >   	 * generated on the port when a cable is not attached.
> >   	 */
> > -	if (IS_G45(dev_priv)) {
> > +	if (IS_G45(i915)) {
> >   		u32 temp = I915_READ(PEG_BAND_GAP_DATA);
> >   		I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd);
> >   	}
> > @@ -3119,7 +3119,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   		DRM_DEBUG_KMS("CEC notifier get failed\n");
> >   }
> > -void intel_hdmi_init(struct drm_i915_private *dev_priv,
> > +void intel_hdmi_init(struct drm_i915_private *i915,
> >   		     i915_reg_t hdmi_reg, enum port port)
> >   {
> >   	struct intel_digital_port *intel_dig_port;
> > @@ -3138,13 +3138,13 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder = &intel_dig_port->base;
> > -	drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> > +	drm_encoder_init(&i915->drm, &intel_encoder->base,
> >   			 &intel_hdmi_enc_funcs, DRM_MODE_ENCODER_TMDS,
> >   			 "HDMI %c", port_name(port));
> >   	intel_encoder->hotplug = intel_encoder_hotplug;
> >   	intel_encoder->compute_config = intel_hdmi_compute_config;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		intel_encoder->disable = pch_disable_hdmi;
> >   		intel_encoder->post_disable = pch_post_disable_hdmi;
> >   	} else {
> > @@ -3152,22 +3152,22 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> >   	}
> >   	intel_encoder->get_hw_state = intel_hdmi_get_hw_state;
> >   	intel_encoder->get_config = intel_hdmi_get_config;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		intel_encoder->pre_pll_enable = chv_hdmi_pre_pll_enable;
> >   		intel_encoder->pre_enable = chv_hdmi_pre_enable;
> >   		intel_encoder->enable = vlv_enable_hdmi;
> >   		intel_encoder->post_disable = chv_hdmi_post_disable;
> >   		intel_encoder->post_pll_disable = chv_hdmi_post_pll_disable;
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915)) {
> >   		intel_encoder->pre_pll_enable = vlv_hdmi_pre_pll_enable;
> >   		intel_encoder->pre_enable = vlv_hdmi_pre_enable;
> >   		intel_encoder->enable = vlv_enable_hdmi;
> >   		intel_encoder->post_disable = vlv_hdmi_post_disable;
> >   	} else {
> >   		intel_encoder->pre_enable = intel_hdmi_pre_enable;
> > -		if (HAS_PCH_CPT(dev_priv))
> > +		if (HAS_PCH_CPT(i915))
> >   			intel_encoder->enable = cpt_enable_hdmi;
> > -		else if (HAS_PCH_IBX(dev_priv))
> > +		else if (HAS_PCH_IBX(i915))
> >   			intel_encoder->enable = ibx_enable_hdmi;
> >   		else
> >   			intel_encoder->enable = g4x_enable_hdmi;
> > @@ -3176,7 +3176,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder->type = INTEL_OUTPUT_HDMI;
> >   	intel_encoder->power_domain = intel_port_to_power_domain(port);
> >   	intel_encoder->port = port;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		if (port == PORT_D)
> >   			intel_encoder->crtc_mask = 1 << 2;
> >   		else
> > @@ -3190,7 +3190,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> >   	 * to work on real hardware. And since g4x can send infoframes to
> >   	 * only one port anyway, nothing is lost by allowing it.
> >   	 */
> > -	if (IS_G4X(dev_priv))
> > +	if (IS_G4X(i915))
> >   		intel_encoder->cloneable |= 1 << INTEL_OUTPUT_HDMI;
> >   	intel_dig_port->hdmi.hdmi_reg = hdmi_reg;
> > @@ -3199,6 +3199,6 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv,
> >   	intel_infoframe_init(intel_dig_port);
> > -	intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port);
> > +	intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port);
> >   	intel_hdmi_init_connector(intel_dig_port, intel_connector);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_hdmi.h b/drivers/gpu/drm/i915/intel_hdmi.h
> > index 106c2e0bc3c9..46f03356b54b 100644
> > --- a/drivers/gpu/drm/i915/intel_hdmi.h
> > +++ b/drivers/gpu/drm/i915/intel_hdmi.h
> > @@ -24,7 +24,7 @@ struct intel_hdmi;
> >   struct drm_connector_state;
> >   union hdmi_infoframe;
> > -void intel_hdmi_init(struct drm_i915_private *dev_priv, i915_reg_t hdmi_reg,
> > +void intel_hdmi_init(struct drm_i915_private *i915, i915_reg_t hdmi_reg,
> >   		     enum port port);
> >   void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port,
> >   			       struct intel_connector *intel_connector);
> > diff --git a/drivers/gpu/drm/i915/intel_hotplug.c b/drivers/gpu/drm/i915/intel_hotplug.c
> > index ff9eb3c855d3..a551b9ab2e17 100644
> > --- a/drivers/gpu/drm/i915/intel_hotplug.c
> > +++ b/drivers/gpu/drm/i915/intel_hotplug.c
> > @@ -78,7 +78,7 @@
> >   /**
> >    * intel_hpd_pin_default - return default pin associated with certain port.
> > - * @dev_priv: private driver data pointer
> > + * @i915: private driver data pointer
> >    * @port: the hpd port to get associated pin
> >    *
> >    * It is only valid and used by digital port encoder.
> > @@ -86,7 +86,7 @@
> >    * Return pin that is associatade with @port and HDP_NONE if no pin is
> >    * hard associated with that @port.
> >    */
> > -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> > +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915,
> >   				   enum port port)
> >   {
> >   	switch (port) {
> > @@ -101,7 +101,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> >   	case PORT_E:
> >   		return HPD_PORT_E;
> >   	case PORT_F:
> > -		if (IS_CNL_WITH_PORT_F(dev_priv))
> > +		if (IS_CNL_WITH_PORT_F(i915))
> >   			return HPD_PORT_E;
> >   		return HPD_PORT_F;
> >   	default:
> > @@ -115,7 +115,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> >   /**
> >    * intel_hpd_irq_storm_detect - gather stats and detect HPD IRQ storm on a pin
> > - * @dev_priv: private driver data pointer
> > + * @i915: private driver data pointer
> >    * @pin: the pin to gather stats on
> >    * @long_hpd: whether the HPD IRQ was long or short
> >    *
> > @@ -124,13 +124,13 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> >    * responsible for further action.
> >    *
> >    * The number of IRQs that are allowed within @HPD_STORM_DETECT_PERIOD is
> > - * stored in @dev_priv->hotplug.hpd_storm_threshold which defaults to
> > + * stored in @i915->hotplug.hpd_storm_threshold which defaults to
> >    * @HPD_STORM_DEFAULT_THRESHOLD. Long IRQs count as +10 to this threshold, and
> >    * short IRQs count as +1. If this threshold is exceeded, it's considered an
> >    * IRQ storm and the IRQ state is set to @HPD_MARK_DISABLED.
> >    *
> >    * By default, most systems will only count long IRQs towards
> > - * &dev_priv->hotplug.hpd_storm_threshold. However, some older systems also
> > + * &i915->hotplug.hpd_storm_threshold. However, some older systems also
> >    * suffer from short IRQ storms and must also track these. Because short IRQ
> >    * storms are naturally caused by sideband interactions with DP MST devices,
> >    * short IRQ detection is only enabled for systems without DP MST support.
> > @@ -142,10 +142,10 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> >    *
> >    * Return true if an IRQ storm was detected on @pin.
> >    */
> > -static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> > +static bool intel_hpd_irq_storm_detect(struct drm_i915_private *i915,
> >   				       enum hpd_pin pin, bool long_hpd)
> >   {
> > -	struct i915_hotplug *hpd = &dev_priv->hotplug;
> > +	struct i915_hotplug *hpd = &i915->hotplug;
> >   	unsigned long start = hpd->stats[pin].last_jiffies;
> >   	unsigned long end = start + msecs_to_jiffies(HPD_STORM_DETECT_PERIOD);
> >   	const int increment = long_hpd ? 10 : 1;
> > @@ -153,7 +153,7 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> >   	bool storm = false;
> >   	if (!threshold ||
> > -	    (!long_hpd && !dev_priv->hotplug.hpd_short_storm_enabled))
> > +	    (!long_hpd && !i915->hotplug.hpd_short_storm_enabled))
> >   		return false;
> >   	if (!time_in_range(jiffies, start, end)) {
> > @@ -175,9 +175,9 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> > +intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_connector *intel_connector;
> >   	struct intel_encoder *intel_encoder;
> >   	struct drm_connector *connector;
> > @@ -185,7 +185,7 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> >   	enum hpd_pin pin;
> >   	bool hpd_disabled = false;
> > -	lockdep_assert_held(&dev_priv->irq_lock);
> > +	lockdep_assert_held(&i915->irq_lock);
> >   	drm_connector_list_iter_begin(dev, &conn_iter);
> >   	drm_for_each_connector_iter(connector, &conn_iter) {
> > @@ -199,14 +199,14 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> >   		pin = intel_encoder->hpd_pin;
> >   		if (pin == HPD_NONE ||
> > -		    dev_priv->hotplug.stats[pin].state != HPD_MARK_DISABLED)
> > +		    i915->hotplug.stats[pin].state != HPD_MARK_DISABLED)
> >   			continue;
> >   		DRM_INFO("HPD interrupt storm detected on connector %s: "
> >   			 "switching from hotplug detection to polling\n",
> >   			 connector->name);
> > -		dev_priv->hotplug.stats[pin].state = HPD_DISABLED;
> > +		i915->hotplug.stats[pin].state = HPD_DISABLED;
> >   		connector->polled = DRM_CONNECTOR_POLL_CONNECT
> >   			| DRM_CONNECTOR_POLL_DISCONNECT;
> >   		hpd_disabled = true;
> > @@ -216,31 +216,31 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv)
> >   	/* Enable polling and queue hotplug re-enabling. */
> >   	if (hpd_disabled) {
> >   		drm_kms_helper_poll_enable(dev);
> > -		mod_delayed_work(system_wq, &dev_priv->hotplug.reenable_work,
> > +		mod_delayed_work(system_wq, &i915->hotplug.reenable_work,
> >   				 msecs_to_jiffies(HPD_STORM_REENABLE_DELAY));
> >   	}
> >   }
> >   static void intel_hpd_irq_storm_reenable_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv),
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915),
> >   			     hotplug.reenable_work.work);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	intel_wakeref_t wakeref;
> >   	enum hpd_pin pin;
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> >   	for_each_hpd_pin(pin) {
> >   		struct drm_connector *connector;
> >   		struct drm_connector_list_iter conn_iter;
> > -		if (dev_priv->hotplug.stats[pin].state != HPD_DISABLED)
> > +		if (i915->hotplug.stats[pin].state != HPD_DISABLED)
> >   			continue;
> > -		dev_priv->hotplug.stats[pin].state = HPD_ENABLED;
> > +		i915->hotplug.stats[pin].state = HPD_ENABLED;
> >   		drm_connector_list_iter_begin(dev, &conn_iter);
> >   		drm_for_each_connector_iter(connector, &conn_iter) {
> > @@ -259,11 +259,11 @@ static void intel_hpd_irq_storm_reenable_work(struct work_struct *work)
> >   		}
> >   		drm_connector_list_iter_end(&conn_iter);
> >   	}
> > -	if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup)
> > -		dev_priv->display.hpd_irq_setup(dev_priv);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	if (i915->display_irqs_enabled && i915->display.hpd_irq_setup)
> > +		i915->display.hpd_irq_setup(i915);
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > +	intel_runtime_pm_put(i915, wakeref);
> >   }
> >   bool intel_encoder_hotplug(struct intel_encoder *encoder,
> > @@ -298,20 +298,20 @@ static bool intel_encoder_has_hpd_pulse(struct intel_encoder *encoder)
> >   static void i915_digport_work_func(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private, hotplug.dig_port_work);
> >   	u32 long_port_mask, short_port_mask;
> >   	struct intel_encoder *encoder;
> >   	u32 old_bits = 0;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	long_port_mask = dev_priv->hotplug.long_port_mask;
> > -	dev_priv->hotplug.long_port_mask = 0;
> > -	short_port_mask = dev_priv->hotplug.short_port_mask;
> > -	dev_priv->hotplug.short_port_mask = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	long_port_mask = i915->hotplug.long_port_mask;
> > +	i915->hotplug.long_port_mask = 0;
> > +	short_port_mask = i915->hotplug.short_port_mask;
> > +	i915->hotplug.short_port_mask = 0;
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		struct intel_digital_port *dig_port;
> >   		enum port port = encoder->port;
> >   		bool long_hpd, short_hpd;
> > @@ -336,10 +336,10 @@ static void i915_digport_work_func(struct work_struct *work)
> >   	}
> >   	if (old_bits) {
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		dev_priv->hotplug.event_bits |= old_bits;
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > -		schedule_work(&dev_priv->hotplug.hotplug_work);
> > +		spin_lock_irq(&i915->irq_lock);
> > +		i915->hotplug.event_bits |= old_bits;
> > +		spin_unlock_irq(&i915->irq_lock);
> > +		schedule_work(&i915->hotplug.hotplug_work);
> >   	}
> >   }
> > @@ -348,9 +348,9 @@ static void i915_digport_work_func(struct work_struct *work)
> >    */
> >   static void i915_hotplug_work_func(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private, hotplug.hotplug_work);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_connector *intel_connector;
> >   	struct intel_encoder *intel_encoder;
> >   	struct drm_connector *connector;
> > @@ -361,15 +361,15 @@ static void i915_hotplug_work_func(struct work_struct *work)
> >   	mutex_lock(&dev->mode_config.mutex);
> >   	DRM_DEBUG_KMS("running encoder hotplug functions\n");
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	hpd_event_bits = dev_priv->hotplug.event_bits;
> > -	dev_priv->hotplug.event_bits = 0;
> > +	hpd_event_bits = i915->hotplug.event_bits;
> > +	i915->hotplug.event_bits = 0;
> >   	/* Enable polling for connectors which had HPD IRQ storms */
> > -	intel_hpd_irq_storm_switch_to_polling(dev_priv);
> > +	intel_hpd_irq_storm_switch_to_polling(i915);
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	drm_connector_list_iter_begin(dev, &conn_iter);
> >   	drm_for_each_connector_iter(connector, &conn_iter) {
> > @@ -395,7 +395,7 @@ static void i915_hotplug_work_func(struct work_struct *work)
> >   /**
> >    * intel_hpd_irq_handler - main hotplug irq handler
> > - * @dev_priv: drm_i915_private
> > + * @i915: drm_i915_private
> >    * @pin_mask: a mask of hpd pins that have triggered the irq
> >    * @long_mask: a mask of hpd pins that may be long hpd pulses
> >    *
> > @@ -409,7 +409,7 @@ static void i915_hotplug_work_func(struct work_struct *work)
> >    * Here, we do hotplug irq storm detection and mitigation, and pass further
> >    * processing to appropriate bottom halves.
> >    */
> > -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_hpd_irq_handler(struct drm_i915_private *i915,
> >   			   u32 pin_mask, u32 long_mask)
> >   {
> >   	struct intel_encoder *encoder;
> > @@ -422,7 +422,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	if (!pin_mask)
> >   		return;
> > -	spin_lock(&dev_priv->irq_lock);
> > +	spin_lock(&i915->irq_lock);
> >   	/*
> >   	 * Determine whether ->hpd_pulse() exists for each pin, and
> > @@ -430,7 +430,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	 * as each pin may have up to two encoders (HDMI and DP) and
> >   	 * only the one of them (DP) will have ->hpd_pulse().
> >   	 */
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		bool has_hpd_pulse = intel_encoder_has_hpd_pulse(encoder);
> >   		enum port port = encoder->port;
> >   		bool long_hpd;
> > @@ -450,10 +450,10 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   		if (long_hpd) {
> >   			long_hpd_pulse_mask |= BIT(pin);
> > -			dev_priv->hotplug.long_port_mask |= BIT(port);
> > +			i915->hotplug.long_port_mask |= BIT(port);
> >   		} else {
> >   			short_hpd_pulse_mask |= BIT(pin);
> > -			dev_priv->hotplug.short_port_mask |= BIT(port);
> > +			i915->hotplug.short_port_mask |= BIT(port);
> >   		}
> >   	}
> > @@ -464,19 +464,19 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   		if (!(BIT(pin) & pin_mask))
> >   			continue;
> > -		if (dev_priv->hotplug.stats[pin].state == HPD_DISABLED) {
> > +		if (i915->hotplug.stats[pin].state == HPD_DISABLED) {
> >   			/*
> >   			 * On GMCH platforms the interrupt mask bits only
> >   			 * prevent irq generation, not the setting of the
> >   			 * hotplug bits itself. So only WARN about unexpected
> >   			 * interrupts on saner platforms.
> >   			 */
> > -			WARN_ONCE(!HAS_GMCH(dev_priv),
> > +			WARN_ONCE(!HAS_GMCH(i915),
> >   				  "Received HPD interrupt on pin %d although disabled\n", pin);
> >   			continue;
> >   		}
> > -		if (dev_priv->hotplug.stats[pin].state != HPD_ENABLED)
> > +		if (i915->hotplug.stats[pin].state != HPD_ENABLED)
> >   			continue;
> >   		/*
> > @@ -487,13 +487,13 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   		if (((short_hpd_pulse_mask | long_hpd_pulse_mask) & BIT(pin))) {
> >   			long_hpd = long_hpd_pulse_mask & BIT(pin);
> >   		} else {
> > -			dev_priv->hotplug.event_bits |= BIT(pin);
> > +			i915->hotplug.event_bits |= BIT(pin);
> >   			long_hpd = true;
> >   			queue_hp = true;
> >   		}
> > -		if (intel_hpd_irq_storm_detect(dev_priv, pin, long_hpd)) {
> > -			dev_priv->hotplug.event_bits &= ~BIT(pin);
> > +		if (intel_hpd_irq_storm_detect(i915, pin, long_hpd)) {
> > +			i915->hotplug.event_bits &= ~BIT(pin);
> >   			storm_detected = true;
> >   			queue_hp = true;
> >   		}
> > @@ -503,9 +503,9 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	 * Disable any IRQs that storms were detected on. Polling enablement
> >   	 * happens later in our hotplug work.
> >   	 */
> > -	if (storm_detected && dev_priv->display_irqs_enabled)
> > -		dev_priv->display.hpd_irq_setup(dev_priv);
> > -	spin_unlock(&dev_priv->irq_lock);
> > +	if (storm_detected && i915->display_irqs_enabled)
> > +		i915->display.hpd_irq_setup(i915);
> > +	spin_unlock(&i915->irq_lock);
> >   	/*
> >   	 * Our hotplug handler can grab modeset locks (by calling down into the
> > @@ -514,14 +514,14 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >   	 * deadlock.
> >   	 */
> >   	if (queue_dig)
> > -		queue_work(dev_priv->hotplug.dp_wq, &dev_priv->hotplug.dig_port_work);
> > +		queue_work(i915->hotplug.dp_wq, &i915->hotplug.dig_port_work);
> >   	if (queue_hp)
> > -		schedule_work(&dev_priv->hotplug.hotplug_work);
> > +		schedule_work(&i915->hotplug.hotplug_work);
> >   }
> >   /**
> >    * intel_hpd_init - initializes and enables hpd support
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function enables the hotplug support. It requires that interrupts have
> >    * already been enabled with intel_irq_init_hw(). From this point on hotplug and
> > @@ -533,43 +533,43 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> >    *
> >    * Also see: intel_hpd_poll_init(), which enables connector polling
> >    */
> > -void intel_hpd_init(struct drm_i915_private *dev_priv)
> > +void intel_hpd_init(struct drm_i915_private *i915)
> >   {
> >   	int i;
> >   	for_each_hpd_pin(i) {
> > -		dev_priv->hotplug.stats[i].count = 0;
> > -		dev_priv->hotplug.stats[i].state = HPD_ENABLED;
> > +		i915->hotplug.stats[i].count = 0;
> > +		i915->hotplug.stats[i].state = HPD_ENABLED;
> >   	}
> > -	WRITE_ONCE(dev_priv->hotplug.poll_enabled, false);
> > -	schedule_work(&dev_priv->hotplug.poll_init_work);
> > +	WRITE_ONCE(i915->hotplug.poll_enabled, false);
> > +	schedule_work(&i915->hotplug.poll_init_work);
> >   	/*
> >   	 * Interrupt setup is already guaranteed to be single-threaded, this is
> >   	 * just to make the assert_spin_locked checks happy.
> >   	 */
> > -	if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup) {
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		if (dev_priv->display_irqs_enabled)
> > -			dev_priv->display.hpd_irq_setup(dev_priv);
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +	if (i915->display_irqs_enabled && i915->display.hpd_irq_setup) {
> > +		spin_lock_irq(&i915->irq_lock);
> > +		if (i915->display_irqs_enabled)
> > +			i915->display.hpd_irq_setup(i915);
> > +		spin_unlock_irq(&i915->irq_lock);
> >   	}
> >   }
> >   static void i915_hpd_poll_init_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(work, struct drm_i915_private,
> >   			     hotplug.poll_init_work);
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	bool enabled;
> >   	mutex_lock(&dev->mode_config.mutex);
> > -	enabled = READ_ONCE(dev_priv->hotplug.poll_enabled);
> > +	enabled = READ_ONCE(i915->hotplug.poll_enabled);
> >   	drm_connector_list_iter_begin(dev, &conn_iter);
> >   	drm_for_each_connector_iter(connector, &conn_iter) {
> > @@ -582,7 +582,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
> >   		if (intel_connector->mst_port)
> >   			continue;
> > -		if (!connector->polled && I915_HAS_HOTPLUG(dev_priv) &&
> > +		if (!connector->polled && I915_HAS_HOTPLUG(i915) &&
> >   		    intel_connector->encoder->hpd_pin > HPD_NONE) {
> >   			connector->polled = enabled ?
> >   				DRM_CONNECTOR_POLL_CONNECT |
> > @@ -607,7 +607,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
> >   /**
> >    * intel_hpd_poll_init - enables/disables polling for connectors with hpd
> > - * @dev_priv: i915 device instance
> > + * @i915: i915 device instance
> >    *
> >    * This function enables polling for all connectors, regardless of whether or
> >    * not they support hotplug detection. Under certain conditions HPD may not be
> > @@ -621,9 +621,9 @@ static void i915_hpd_poll_init_work(struct work_struct *work)
> >    *
> >    * Also see: intel_hpd_init(), which restores hpd handling.
> >    */
> > -void intel_hpd_poll_init(struct drm_i915_private *dev_priv)
> > +void intel_hpd_poll_init(struct drm_i915_private *i915)
> >   {
> > -	WRITE_ONCE(dev_priv->hotplug.poll_enabled, true);
> > +	WRITE_ONCE(i915->hotplug.poll_enabled, true);
> >   	/*
> >   	 * We might already be holding dev->mode_config.mutex, so do this in a
> > @@ -631,57 +631,57 @@ void intel_hpd_poll_init(struct drm_i915_private *dev_priv)
> >   	 * As well, there's no issue if we race here since we always reschedule
> >   	 * this worker anyway
> >   	 */
> > -	schedule_work(&dev_priv->hotplug.poll_init_work);
> > +	schedule_work(&i915->hotplug.poll_init_work);
> >   }
> > -void intel_hpd_init_work(struct drm_i915_private *dev_priv)
> > +void intel_hpd_init_work(struct drm_i915_private *i915)
> >   {
> > -	INIT_WORK(&dev_priv->hotplug.hotplug_work, i915_hotplug_work_func);
> > -	INIT_WORK(&dev_priv->hotplug.dig_port_work, i915_digport_work_func);
> > -	INIT_WORK(&dev_priv->hotplug.poll_init_work, i915_hpd_poll_init_work);
> > -	INIT_DELAYED_WORK(&dev_priv->hotplug.reenable_work,
> > +	INIT_WORK(&i915->hotplug.hotplug_work, i915_hotplug_work_func);
> > +	INIT_WORK(&i915->hotplug.dig_port_work, i915_digport_work_func);
> > +	INIT_WORK(&i915->hotplug.poll_init_work, i915_hpd_poll_init_work);
> > +	INIT_DELAYED_WORK(&i915->hotplug.reenable_work,
> >   			  intel_hpd_irq_storm_reenable_work);
> >   }
> > -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv)
> > +void intel_hpd_cancel_work(struct drm_i915_private *i915)
> >   {
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > -	dev_priv->hotplug.long_port_mask = 0;
> > -	dev_priv->hotplug.short_port_mask = 0;
> > -	dev_priv->hotplug.event_bits = 0;
> > +	i915->hotplug.long_port_mask = 0;
> > +	i915->hotplug.short_port_mask = 0;
> > +	i915->hotplug.event_bits = 0;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> > -	cancel_work_sync(&dev_priv->hotplug.dig_port_work);
> > -	cancel_work_sync(&dev_priv->hotplug.hotplug_work);
> > -	cancel_work_sync(&dev_priv->hotplug.poll_init_work);
> > -	cancel_delayed_work_sync(&dev_priv->hotplug.reenable_work);
> > +	cancel_work_sync(&i915->hotplug.dig_port_work);
> > +	cancel_work_sync(&i915->hotplug.hotplug_work);
> > +	cancel_work_sync(&i915->hotplug.poll_init_work);
> > +	cancel_delayed_work_sync(&i915->hotplug.reenable_work);
> >   }
> > -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin)
> > +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin)
> >   {
> >   	bool ret = false;
> >   	if (pin == HPD_NONE)
> >   		return false;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	if (dev_priv->hotplug.stats[pin].state == HPD_ENABLED) {
> > -		dev_priv->hotplug.stats[pin].state = HPD_DISABLED;
> > +	spin_lock_irq(&i915->irq_lock);
> > +	if (i915->hotplug.stats[pin].state == HPD_ENABLED) {
> > +		i915->hotplug.stats[pin].state = HPD_DISABLED;
> >   		ret = true;
> >   	}
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_unlock_irq(&i915->irq_lock);
> >   	return ret;
> >   }
> > -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin)
> > +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin)
> >   {
> >   	if (pin == HPD_NONE)
> >   		return;
> > -	spin_lock_irq(&dev_priv->irq_lock);
> > -	dev_priv->hotplug.stats[pin].state = HPD_ENABLED;
> > -	spin_unlock_irq(&dev_priv->irq_lock);
> > +	spin_lock_irq(&i915->irq_lock);
> > +	i915->hotplug.stats[pin].state = HPD_ENABLED;
> > +	spin_unlock_irq(&i915->irq_lock);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_hotplug.h b/drivers/gpu/drm/i915/intel_hotplug.h
> > index 805f897dbb7a..2fcc3148b369 100644
> > --- a/drivers/gpu/drm/i915/intel_hotplug.h
> > +++ b/drivers/gpu/drm/i915/intel_hotplug.h
> > @@ -14,17 +14,17 @@ struct drm_i915_private;
> >   struct intel_connector;
> >   struct intel_encoder;
> > -void intel_hpd_poll_init(struct drm_i915_private *dev_priv);
> > +void intel_hpd_poll_init(struct drm_i915_private *i915);
> >   bool intel_encoder_hotplug(struct intel_encoder *encoder,
> >   			   struct intel_connector *connector);
> > -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv,
> > +void intel_hpd_irq_handler(struct drm_i915_private *i915,
> >   			   u32 pin_mask, u32 long_mask);
> > -void intel_hpd_init(struct drm_i915_private *dev_priv);
> > -void intel_hpd_init_work(struct drm_i915_private *dev_priv);
> > -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv);
> > -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv,
> > +void intel_hpd_init(struct drm_i915_private *i915);
> > +void intel_hpd_init_work(struct drm_i915_private *i915);
> > +void intel_hpd_cancel_work(struct drm_i915_private *i915);
> > +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915,
> >   				   enum port port);
> > -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin);
> > -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin);
> > +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin);
> > +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin);
> >   #endif /* __INTEL_HOTPLUG_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_huc.c b/drivers/gpu/drm/i915/intel_huc.c
> > index 8572a0588efc..2b23662698a9 100644
> > --- a/drivers/gpu/drm/i915/intel_huc.c
> > +++ b/drivers/gpu/drm/i915/intel_huc.c
> > @@ -167,14 +167,14 @@ int intel_huc_auth(struct intel_huc *huc)
> >    */
> >   int intel_huc_check_status(struct intel_huc *huc)
> >   {
> > -	struct drm_i915_private *dev_priv = huc_to_i915(huc);
> > +	struct drm_i915_private *i915 = huc_to_i915(huc);
> >   	intel_wakeref_t wakeref;
> >   	bool status = false;
> > -	if (!HAS_HUC(dev_priv))
> > +	if (!HAS_HUC(i915))
> >   		return -ENODEV;
> > -	with_intel_runtime_pm(dev_priv, wakeref)
> > +	with_intel_runtime_pm(i915, wakeref)
> >   		status = (I915_READ(huc->status.reg) & huc->status.mask) ==
> >   			  huc->status.value;
> > diff --git a/drivers/gpu/drm/i915/intel_huc_fw.c b/drivers/gpu/drm/i915/intel_huc_fw.c
> > index 05cbf8338f53..03aed0664726 100644
> > --- a/drivers/gpu/drm/i915/intel_huc_fw.c
> > +++ b/drivers/gpu/drm/i915/intel_huc_fw.c
> > @@ -69,34 +69,34 @@ MODULE_FIRMWARE(I915_ICL_HUC_UCODE);
> >   static void huc_fw_select(struct intel_uc_fw *huc_fw)
> >   {
> >   	struct intel_huc *huc = container_of(huc_fw, struct intel_huc, fw);
> > -	struct drm_i915_private *dev_priv = huc_to_i915(huc);
> > +	struct drm_i915_private *i915 = huc_to_i915(huc);
> >   	GEM_BUG_ON(huc_fw->type != INTEL_UC_FW_TYPE_HUC);
> > -	if (!HAS_HUC(dev_priv))
> > +	if (!HAS_HUC(i915))
> >   		return;
> >   	if (i915_modparams.huc_firmware_path) {
> >   		huc_fw->path = i915_modparams.huc_firmware_path;
> >   		huc_fw->major_ver_wanted = 0;
> >   		huc_fw->minor_ver_wanted = 0;
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)) {
> >   		huc_fw->path = I915_SKL_HUC_UCODE;
> >   		huc_fw->major_ver_wanted = SKL_HUC_FW_MAJOR;
> >   		huc_fw->minor_ver_wanted = SKL_HUC_FW_MINOR;
> > -	} else if (IS_BROXTON(dev_priv)) {
> > +	} else if (IS_BROXTON(i915)) {
> >   		huc_fw->path = I915_BXT_HUC_UCODE;
> >   		huc_fw->major_ver_wanted = BXT_HUC_FW_MAJOR;
> >   		huc_fw->minor_ver_wanted = BXT_HUC_FW_MINOR;
> > -	} else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) {
> > +	} else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) {
> >   		huc_fw->path = I915_KBL_HUC_UCODE;
> >   		huc_fw->major_ver_wanted = KBL_HUC_FW_MAJOR;
> >   		huc_fw->minor_ver_wanted = KBL_HUC_FW_MINOR;
> > -	} else if (IS_GEMINILAKE(dev_priv)) {
> > +	} else if (IS_GEMINILAKE(i915)) {
> >   		huc_fw->path = I915_GLK_HUC_UCODE;
> >   		huc_fw->major_ver_wanted = GLK_HUC_FW_MAJOR;
> >   		huc_fw->minor_ver_wanted = GLK_HUC_FW_MINOR;
> > -	} else if (IS_ICELAKE(dev_priv)) {
> > +	} else if (IS_ICELAKE(i915)) {
> >   		huc_fw->path = I915_ICL_HUC_UCODE;
> >   		huc_fw->major_ver_wanted = ICL_HUC_FW_MAJOR;
> >   		huc_fw->minor_ver_wanted = ICL_HUC_FW_MINOR;
> > @@ -135,8 +135,8 @@ static void huc_xfer_rsa(struct intel_huc *huc)
> >   static int huc_xfer_ucode(struct intel_huc *huc)
> >   {
> >   	struct intel_uc_fw *huc_fw = &huc->fw;
> > -	struct drm_i915_private *dev_priv = huc_to_i915(huc);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = huc_to_i915(huc);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	unsigned long offset = 0;
> >   	u32 size;
> >   	int ret;
> > diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.c b/drivers/gpu/drm/i915/intel_lpe_audio.c
> > index b19800b58442..d95ae05ccebc 100644
> > --- a/drivers/gpu/drm/i915/intel_lpe_audio.c
> > +++ b/drivers/gpu/drm/i915/intel_lpe_audio.c
> > @@ -73,12 +73,12 @@
> >   #include "i915_drv.h"
> >   #include "intel_lpe_audio.h"
> > -#define HAS_LPE_AUDIO(dev_priv) ((dev_priv)->lpe_audio.platdev != NULL)
> > +#define HAS_LPE_AUDIO(i915) ((i915)->lpe_audio.platdev != NULL)
> >   static struct platform_device *
> > -lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> > +lpe_audio_platdev_create(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct platform_device_info pinfo = {};
> >   	struct resource *rsc;
> >   	struct platform_device *platdev;
> > @@ -94,7 +94,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> >   		return ERR_PTR(-ENOMEM);
> >   	}
> > -	rsc[0].start    = rsc[0].end = dev_priv->lpe_audio.irq;
> > +	rsc[0].start    = rsc[0].end = i915->lpe_audio.irq;
> >   	rsc[0].flags    = IORESOURCE_IRQ;
> >   	rsc[0].name     = "hdmi-lpe-audio-irq";
> > @@ -114,8 +114,8 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> >   	pinfo.size_data = sizeof(*pdata);
> >   	pinfo.dma_mask = DMA_BIT_MASK(32);
> > -	pdata->num_pipes = INTEL_INFO(dev_priv)->num_pipes;
> > -	pdata->num_ports = IS_CHERRYVIEW(dev_priv) ? 3 : 2; /* B,C,D or B,C */
> > +	pdata->num_pipes = INTEL_INFO(i915)->num_pipes;
> > +	pdata->num_ports = IS_CHERRYVIEW(i915) ? 3 : 2; /* B,C,D or B,C */
> >   	pdata->port[0].pipe = -1;
> >   	pdata->port[1].pipe = -1;
> >   	pdata->port[2].pipe = -1;
> > @@ -135,7 +135,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv)
> >   	return platdev;
> >   }
> > -static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv)
> > +static void lpe_audio_platdev_destroy(struct drm_i915_private *i915)
> >   {
> >   	/* XXX Note that platform_device_register_full() allocates a dma_mask
> >   	 * and never frees it. We can't free it here as we cannot guarantee
> > @@ -145,7 +145,7 @@ static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv)
> >   	 * than us fiddle with its internals.
> >   	 */
> > -	platform_device_unregister(dev_priv->lpe_audio.platdev);
> > +	platform_device_unregister(i915->lpe_audio.platdev);
> >   }
> >   static void lpe_audio_irq_unmask(struct irq_data *d)
> > @@ -162,24 +162,24 @@ static struct irq_chip lpe_audio_irqchip = {
> >   	.irq_unmask = lpe_audio_irq_unmask,
> >   };
> > -static int lpe_audio_irq_init(struct drm_i915_private *dev_priv)
> > +static int lpe_audio_irq_init(struct drm_i915_private *i915)
> >   {
> > -	int irq = dev_priv->lpe_audio.irq;
> > +	int irq = i915->lpe_audio.irq;
> > -	WARN_ON(!intel_irqs_enabled(dev_priv));
> > +	WARN_ON(!intel_irqs_enabled(i915));
> >   	irq_set_chip_and_handler_name(irq,
> >   				&lpe_audio_irqchip,
> >   				handle_simple_irq,
> >   				"hdmi_lpe_audio_irq_handler");
> > -	return irq_set_chip_data(irq, dev_priv);
> > +	return irq_set_chip_data(irq, i915);
> >   }
> > -static bool lpe_audio_detect(struct drm_i915_private *dev_priv)
> > +static bool lpe_audio_detect(struct drm_i915_private *i915)
> >   {
> >   	int lpe_present = false;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		static const struct pci_device_id atom_hdaudio_ids[] = {
> >   			/* Baytrail */
> >   			{PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f04)},
> > @@ -196,21 +196,21 @@ static bool lpe_audio_detect(struct drm_i915_private *dev_priv)
> >   	return lpe_present;
> >   }
> > -static int lpe_audio_setup(struct drm_i915_private *dev_priv)
> > +static int lpe_audio_setup(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	dev_priv->lpe_audio.irq = irq_alloc_desc(0);
> > -	if (dev_priv->lpe_audio.irq < 0) {
> > +	i915->lpe_audio.irq = irq_alloc_desc(0);
> > +	if (i915->lpe_audio.irq < 0) {
> >   		DRM_ERROR("Failed to allocate IRQ desc: %d\n",
> > -			dev_priv->lpe_audio.irq);
> > -		ret = dev_priv->lpe_audio.irq;
> > +			i915->lpe_audio.irq);
> > +		ret = i915->lpe_audio.irq;
> >   		goto err;
> >   	}
> > -	DRM_DEBUG("irq = %d\n", dev_priv->lpe_audio.irq);
> > +	DRM_DEBUG("irq = %d\n", i915->lpe_audio.irq);
> > -	ret = lpe_audio_irq_init(dev_priv);
> > +	ret = lpe_audio_irq_init(i915);
> >   	if (ret) {
> >   		DRM_ERROR("Failed to initialize irqchip for lpe audio: %d\n",
> > @@ -218,10 +218,10 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv)
> >   		goto err_free_irq;
> >   	}
> > -	dev_priv->lpe_audio.platdev = lpe_audio_platdev_create(dev_priv);
> > +	i915->lpe_audio.platdev = lpe_audio_platdev_create(i915);
> > -	if (IS_ERR(dev_priv->lpe_audio.platdev)) {
> > -		ret = PTR_ERR(dev_priv->lpe_audio.platdev);
> > +	if (IS_ERR(i915->lpe_audio.platdev)) {
> > +		ret = PTR_ERR(i915->lpe_audio.platdev);
> >   		DRM_ERROR("Failed to create lpe audio platform device: %d\n",
> >   			ret);
> >   		goto err_free_irq;
> > @@ -234,28 +234,28 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv)
> >   	return 0;
> >   err_free_irq:
> > -	irq_free_desc(dev_priv->lpe_audio.irq);
> > +	irq_free_desc(i915->lpe_audio.irq);
> >   err:
> > -	dev_priv->lpe_audio.irq = -1;
> > -	dev_priv->lpe_audio.platdev = NULL;
> > +	i915->lpe_audio.irq = -1;
> > +	i915->lpe_audio.platdev = NULL;
> >   	return ret;
> >   }
> >   /**
> >    * intel_lpe_audio_irq_handler() - forwards the LPE audio irq
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    *
> >    * the LPE Audio irq is forwarded to the irq handler registered by LPE audio
> >    * driver.
> >    */
> > -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv)
> > +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	if (!HAS_LPE_AUDIO(dev_priv))
> > +	if (!HAS_LPE_AUDIO(i915))
> >   		return;
> > -	ret = generic_handle_irq(dev_priv->lpe_audio.irq);
> > +	ret = generic_handle_irq(i915->lpe_audio.irq);
> >   	if (ret)
> >   		DRM_ERROR_RATELIMITED("error handling LPE audio irq: %d\n",
> >   				ret);
> > @@ -264,17 +264,17 @@ void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_lpe_audio_init() - detect and setup the bridge between HDMI LPE Audio
> >    * driver and i915
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    *
> >    * Return: 0 if successful. non-zero if detection or
> >    * llocation/initialization fails
> >    */
> > -int intel_lpe_audio_init(struct drm_i915_private *dev_priv)
> > +int intel_lpe_audio_init(struct drm_i915_private *i915)
> >   {
> >   	int ret = -ENODEV;
> > -	if (lpe_audio_detect(dev_priv)) {
> > -		ret = lpe_audio_setup(dev_priv);
> > +	if (lpe_audio_detect(i915)) {
> > +		ret = lpe_audio_setup(i915);
> >   		if (ret < 0)
> >   			DRM_ERROR("failed to setup LPE Audio bridge\n");
> >   	}
> > @@ -284,31 +284,31 @@ int intel_lpe_audio_init(struct drm_i915_private *dev_priv)
> >   /**
> >    * intel_lpe_audio_teardown() - destroy the bridge between HDMI LPE
> >    * audio driver and i915
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    *
> >    * release all the resources for LPE audio <-> i915 bridge.
> >    */
> > -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv)
> > +void intel_lpe_audio_teardown(struct drm_i915_private *i915)
> >   {
> >   	struct irq_desc *desc;
> > -	if (!HAS_LPE_AUDIO(dev_priv))
> > +	if (!HAS_LPE_AUDIO(i915))
> >   		return;
> > -	desc = irq_to_desc(dev_priv->lpe_audio.irq);
> > +	desc = irq_to_desc(i915->lpe_audio.irq);
> > -	lpe_audio_platdev_destroy(dev_priv);
> > +	lpe_audio_platdev_destroy(i915);
> > -	irq_free_desc(dev_priv->lpe_audio.irq);
> > +	irq_free_desc(i915->lpe_audio.irq);
> > -	dev_priv->lpe_audio.irq = -1;
> > -	dev_priv->lpe_audio.platdev = NULL;
> > +	i915->lpe_audio.irq = -1;
> > +	i915->lpe_audio.platdev = NULL;
> >   }
> >   /**
> >    * intel_lpe_audio_notify() - notify lpe audio event
> >    * audio driver and i915
> > - * @dev_priv: the i915 drm device private data
> > + * @i915: the i915 drm device private data
> >    * @pipe: pipe
> >    * @port: port
> >    * @eld : ELD data
> > @@ -317,7 +317,7 @@ void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv)
> >    *
> >    * Notify lpe audio driver of eld change.
> >    */
> > -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> > +void intel_lpe_audio_notify(struct drm_i915_private *i915,
> >   			    enum pipe pipe, enum port port,
> >   			    const void *eld, int ls_clock, bool dp_output)
> >   {
> > @@ -326,10 +326,10 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> >   	struct intel_hdmi_lpe_audio_port_pdata *ppdata;
> >   	u32 audio_enable;
> > -	if (!HAS_LPE_AUDIO(dev_priv))
> > +	if (!HAS_LPE_AUDIO(i915))
> >   		return;
> > -	pdata = dev_get_platdata(&dev_priv->lpe_audio.platdev->dev);
> > +	pdata = dev_get_platdata(&i915->lpe_audio.platdev->dev);
> >   	ppdata = &pdata->port[port - PORT_B];
> >   	spin_lock_irqsave(&pdata->lpe_audio_slock, irqflags);
> > @@ -357,7 +357,7 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> >   	}
> >   	if (pdata->notify_audio_lpe)
> > -		pdata->notify_audio_lpe(dev_priv->lpe_audio.platdev, port - PORT_B);
> > +		pdata->notify_audio_lpe(i915->lpe_audio.platdev, port - PORT_B);
> >   	spin_unlock_irqrestore(&pdata->lpe_audio_slock, irqflags);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.h b/drivers/gpu/drm/i915/intel_lpe_audio.h
> > index f848c5038714..df43abce3c95 100644
> > --- a/drivers/gpu/drm/i915/intel_lpe_audio.h
> > +++ b/drivers/gpu/drm/i915/intel_lpe_audio.h
> > @@ -12,10 +12,10 @@ enum pipe;
> >   enum port;
> >   struct drm_i915_private;
> > -int  intel_lpe_audio_init(struct drm_i915_private *dev_priv);
> > -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv);
> > -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv);
> > -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv,
> > +int  intel_lpe_audio_init(struct drm_i915_private *i915);
> > +void intel_lpe_audio_teardown(struct drm_i915_private *i915);
> > +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915);
> > +void intel_lpe_audio_notify(struct drm_i915_private *i915,
> >   			    enum pipe pipe, enum port port,
> >   			    const void *eld, int ls_clock, bool dp_output);
> > diff --git a/drivers/gpu/drm/i915/intel_lspcon.c b/drivers/gpu/drm/i915/intel_lspcon.c
> > index 7028d0cf3bb1..695465f97195 100644
> > --- a/drivers/gpu/drm/i915/intel_lspcon.c
> > +++ b/drivers/gpu/drm/i915/intel_lspcon.c
> > @@ -555,10 +555,10 @@ bool lspcon_init(struct intel_digital_port *intel_dig_port)
> >   	struct intel_dp *dp = &intel_dig_port->dp;
> >   	struct intel_lspcon *lspcon = &intel_dig_port->lspcon;
> >   	struct drm_device *dev = intel_dig_port->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_connector *connector = &dp->attached_connector->base;
> > -	if (!HAS_LSPCON(dev_priv)) {
> > +	if (!HAS_LSPCON(i915)) {
> >   		DRM_ERROR("LSPCON is not supported on this platform\n");
> >   		return false;
> >   	}
> > diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c
> > index efefed62a7f8..0405f9834827 100644
> > --- a/drivers/gpu/drm/i915/intel_lvds.c
> > +++ b/drivers/gpu/drm/i915/intel_lvds.c
> > @@ -80,7 +80,7 @@ static struct intel_lvds_encoder *to_lvds_encoder(struct drm_encoder *encoder)
> >   	return container_of(encoder, struct intel_lvds_encoder, base.base);
> >   }
> > -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_lvds_port_enabled(struct drm_i915_private *i915,
> >   			     i915_reg_t lvds_reg, enum pipe *pipe)
> >   {
> >   	u32 val;
> > @@ -88,7 +88,7 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(lvds_reg);
> >   	/* asserts want to know the pipe even if the port is disabled */
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		*pipe = (val & LVDS_PIPE_SEL_MASK_CPT) >> LVDS_PIPE_SEL_SHIFT_CPT;
> >   	else
> >   		*pipe = (val & LVDS_PIPE_SEL_MASK) >> LVDS_PIPE_SEL_SHIFT;
> > @@ -99,19 +99,19 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> >   static bool intel_lvds_get_hw_state(struct intel_encoder *encoder,
> >   				    enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > -	ret = intel_lvds_port_enabled(dev_priv, lvds_encoder->reg, pipe);
> > +	ret = intel_lvds_port_enabled(i915, lvds_encoder->reg, pipe);
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -119,7 +119,7 @@ static bool intel_lvds_get_hw_state(struct intel_encoder *encoder,
> >   static void intel_lvds_get_config(struct intel_encoder *encoder,
> >   				  struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> >   	u32 tmp, flags = 0;
> > @@ -137,12 +137,12 @@ static void intel_lvds_get_config(struct intel_encoder *encoder,
> >   	pipe_config->base.adjusted_mode.flags |= flags;
> > -	if (INTEL_GEN(dev_priv) < 5)
> > +	if (INTEL_GEN(i915) < 5)
> >   		pipe_config->gmch_pfit.lvds_border_bits =
> >   			tmp & LVDS_BORDER_ENABLE;
> >   	/* gen2/3 store dither state in pfit control, needs to match */
> > -	if (INTEL_GEN(dev_priv) < 4) {
> > +	if (INTEL_GEN(i915) < 4) {
> >   		tmp = I915_READ(PFIT_CONTROL);
> >   		pipe_config->gmch_pfit.control |= tmp & PANEL_8TO6_DITHER_ENABLE;
> > @@ -151,7 +151,7 @@ static void intel_lvds_get_config(struct intel_encoder *encoder,
> >   	pipe_config->base.adjusted_mode.crtc_clock = pipe_config->port_clock;
> >   }
> > -static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> > +static void intel_lvds_pps_get_hw_state(struct drm_i915_private *i915,
> >   					struct intel_lvds_pps *pps)
> >   {
> >   	u32 val;
> > @@ -180,7 +180,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> >   	/* Convert from 100ms to 100us units */
> >   	pps->t4 = val * 1000;
> > -	if (INTEL_GEN(dev_priv) <= 4 &&
> > +	if (INTEL_GEN(i915) <= 4 &&
> >   	    pps->t1_t2 == 0 && pps->t5 == 0 && pps->t3 == 0 && pps->tx == 0) {
> >   		DRM_DEBUG_KMS("Panel power timings uninitialized, "
> >   			      "setting defaults\n");
> > @@ -198,7 +198,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv,
> >   			 pps->divider, pps->port, pps->powerdown_on_reset);
> >   }
> > -static void intel_lvds_pps_init_hw(struct drm_i915_private *dev_priv,
> > +static void intel_lvds_pps_init_hw(struct drm_i915_private *i915,
> >   				   struct intel_lvds_pps *pps)
> >   {
> >   	u32 val;
> > @@ -229,26 +229,26 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> >   	int pipe = crtc->pipe;
> >   	u32 temp;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > -		assert_fdi_rx_pll_disabled(dev_priv, pipe);
> > -		assert_shared_dpll_disabled(dev_priv,
> > +	if (HAS_PCH_SPLIT(i915)) {
> > +		assert_fdi_rx_pll_disabled(i915, pipe);
> > +		assert_shared_dpll_disabled(i915,
> >   					    pipe_config->shared_dpll);
> >   	} else {
> > -		assert_pll_disabled(dev_priv, pipe);
> > +		assert_pll_disabled(i915, pipe);
> >   	}
> > -	intel_lvds_pps_init_hw(dev_priv, &lvds_encoder->init_pps);
> > +	intel_lvds_pps_init_hw(i915, &lvds_encoder->init_pps);
> >   	temp = lvds_encoder->init_lvds_val;
> >   	temp |= LVDS_PORT_EN | LVDS_A0A2_CLKA_POWER_UP;
> > -	if (HAS_PCH_CPT(dev_priv)) {
> > +	if (HAS_PCH_CPT(i915)) {
> >   		temp &= ~LVDS_PIPE_SEL_MASK_CPT;
> >   		temp |= LVDS_PIPE_SEL_CPT(pipe);
> >   	} else {
> > @@ -283,7 +283,7 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder,
> >   	 * special lvds dither control bit on pch-split platforms, dithering is
> >   	 * only controlled through the PIPECONF reg.
> >   	 */
> > -	if (IS_GEN(dev_priv, 4)) {
> > +	if (IS_GEN(i915, 4)) {
> >   		/*
> >   		 * Bspec wording suggests that LVDS port dithering only exists
> >   		 * for 18bpp panels.
> > @@ -311,14 +311,14 @@ static void intel_enable_lvds(struct intel_encoder *encoder,
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> >   	struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	I915_WRITE(lvds_encoder->reg, I915_READ(lvds_encoder->reg) | LVDS_PORT_EN);
> >   	I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) | PANEL_POWER_ON);
> >   	POSTING_READ(lvds_encoder->reg);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    PP_STATUS(0), PP_ON, PP_ON, 5000))
> >   		DRM_ERROR("timed out waiting for panel to power on\n");
> > @@ -330,10 +330,10 @@ static void intel_disable_lvds(struct intel_encoder *encoder,
> >   			       const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) & ~PANEL_POWER_ON);
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    PP_STATUS(0), PP_ON, 0, 1000))
> >   		DRM_ERROR("timed out waiting for panel to power off\n");
> > @@ -389,7 +389,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> >   				     struct intel_crtc_state *pipe_config,
> >   				     struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> >   	struct intel_lvds_encoder *lvds_encoder =
> >   		to_lvds_encoder(&intel_encoder->base);
> >   	struct intel_connector *intel_connector =
> > @@ -399,7 +399,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> >   	unsigned int lvds_bpp;
> >   	/* Should never happen!! */
> > -	if (INTEL_GEN(dev_priv) < 4 && intel_crtc->pipe == 0) {
> > +	if (INTEL_GEN(i915) < 4 && intel_crtc->pipe == 0) {
> >   		DRM_ERROR("Can't support LVDS on pipe A\n");
> >   		return -EINVAL;
> >   	}
> > @@ -429,7 +429,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder,
> >   	if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN)
> >   		return -EINVAL;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		pipe_config->has_pch_encoder = true;
> >   		intel_pch_panel_fitting(intel_crtc, pipe_config,
> > @@ -751,11 +751,11 @@ static const struct dmi_system_id intel_dual_link_lvds[] = {
> >   	{ }	/* terminating entry */
> >   };
> > -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv)
> > +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915)
> >   {
> >   	struct intel_encoder *encoder;
> > -	for_each_intel_encoder(&dev_priv->drm, encoder) {
> > +	for_each_intel_encoder(&i915->drm, encoder) {
> >   		if (encoder->type == INTEL_OUTPUT_LVDS)
> >   			return encoder;
> >   	}
> > @@ -763,9 +763,9 @@ struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv)
> >   	return NULL;
> >   }
> > -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv)
> > +bool intel_is_dual_link_lvds(struct drm_i915_private *i915)
> >   {
> > -	struct intel_encoder *encoder = intel_get_lvds_encoder(dev_priv);
> > +	struct intel_encoder *encoder = intel_get_lvds_encoder(i915);
> >   	return encoder && to_lvds_encoder(&encoder->base)->is_dual_link;
> >   }
> > @@ -774,7 +774,7 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder)
> >   {
> >   	struct drm_device *dev = lvds_encoder->base.base.dev;
> >   	unsigned int val;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	/* use the module option value if specified */
> >   	if (i915_modparams.lvds_channel_mode > 0)
> > @@ -794,26 +794,26 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder)
> >   	 * register is uninitialized.
> >   	 */
> >   	val = I915_READ(lvds_encoder->reg);
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK_CPT);
> >   	else
> >   		val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK);
> >   	if (val == 0)
> > -		val = dev_priv->vbt.bios_lvds_val;
> > +		val = i915->vbt.bios_lvds_val;
> >   	return (val & LVDS_CLKB_POWER_MASK) == LVDS_CLKB_POWER_UP;
> >   }
> >   /**
> >    * intel_lvds_init - setup LVDS connectors on this device
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    *
> >    * Create the connector, register the LVDS DDC bus, and try to figure out what
> >    * modes we can display on the LVDS panel (if present).
> >    */
> > -void intel_lvds_init(struct drm_i915_private *dev_priv)
> > +void intel_lvds_init(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_lvds_encoder *lvds_encoder;
> >   	struct intel_encoder *intel_encoder;
> >   	struct intel_connector *intel_connector;
> > @@ -829,30 +829,30 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> >   	/* Skip init on machines we know falsely report LVDS */
> >   	if (dmi_check_system(intel_no_lvds)) {
> > -		WARN(!dev_priv->vbt.int_lvds_support,
> > +		WARN(!i915->vbt.int_lvds_support,
> >   		     "Useless DMI match. Internal LVDS support disabled by VBT\n");
> >   		return;
> >   	}
> > -	if (!dev_priv->vbt.int_lvds_support) {
> > +	if (!i915->vbt.int_lvds_support) {
> >   		DRM_DEBUG_KMS("Internal LVDS support disabled by VBT\n");
> >   		return;
> >   	}
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		lvds_reg = PCH_LVDS;
> >   	else
> >   		lvds_reg = LVDS;
> >   	lvds = I915_READ(lvds_reg);
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		if ((lvds & LVDS_DETECTED) == 0)
> >   			return;
> >   	}
> >   	pin = GMBUS_PIN_PANEL;
> > -	if (!intel_bios_is_lvds_present(dev_priv, &pin)) {
> > +	if (!intel_bios_is_lvds_present(i915, &pin)) {
> >   		if ((lvds & LVDS_PORT_EN) == 0) {
> >   			DRM_DEBUG_KMS("LVDS is not present in VBT\n");
> >   			return;
> > @@ -884,7 +884,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> >   	intel_encoder->enable = intel_enable_lvds;
> >   	intel_encoder->pre_enable = intel_pre_enable_lvds;
> >   	intel_encoder->compute_config = intel_lvds_compute_config;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		intel_encoder->disable = pch_disable_lvds;
> >   		intel_encoder->post_disable = pch_post_disable_lvds;
> >   	} else {
> > @@ -901,9 +901,9 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> >   	intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER;
> >   	intel_encoder->port = PORT_NONE;
> >   	intel_encoder->cloneable = 0;
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		intel_encoder->crtc_mask = (1 << 0) | (1 << 1) | (1 << 2);
> > -	else if (IS_GEN(dev_priv, 4))
> > +	else if (IS_GEN(i915, 4))
> >   		intel_encoder->crtc_mask = (1 << 0) | (1 << 1);
> >   	else
> >   		intel_encoder->crtc_mask = (1 << 1);
> > @@ -922,7 +922,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> >   	drm_connector_attach_scaling_mode_property(connector, allowed_scalers);
> >   	connector->state->scaling_mode = DRM_MODE_SCALE_ASPECT;
> > -	intel_lvds_pps_get_hw_state(dev_priv, &lvds_encoder->init_pps);
> > +	intel_lvds_pps_get_hw_state(i915, &lvds_encoder->init_pps);
> >   	lvds_encoder->init_lvds_val = lvds;
> >   	/*
> > @@ -940,10 +940,10 @@ void intel_lvds_init(struct drm_i915_private *dev_priv)
> >   	mutex_lock(&dev->mode_config.mutex);
> >   	if (vga_switcheroo_handler_flags() & VGA_SWITCHEROO_CAN_SWITCH_DDC)
> >   		edid = drm_get_edid_switcheroo(connector,
> > -				    intel_gmbus_get_adapter(dev_priv, pin));
> > +				    intel_gmbus_get_adapter(i915, pin));
> >   	else
> >   		edid = drm_get_edid(connector,
> > -				    intel_gmbus_get_adapter(dev_priv, pin));
> > +				    intel_gmbus_get_adapter(i915, pin));
> >   	if (edid) {
> >   		if (drm_add_edid_modes(connector, edid)) {
> >   			drm_connector_update_edid_property(connector,
> > diff --git a/drivers/gpu/drm/i915/intel_lvds.h b/drivers/gpu/drm/i915/intel_lvds.h
> > index bc9c8b84ba2f..4afdcb4d6352 100644
> > --- a/drivers/gpu/drm/i915/intel_lvds.h
> > +++ b/drivers/gpu/drm/i915/intel_lvds.h
> > @@ -13,10 +13,10 @@
> >   enum pipe;
> >   struct drm_i915_private;
> > -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_lvds_port_enabled(struct drm_i915_private *i915,
> >   			     i915_reg_t lvds_reg, enum pipe *pipe);
> > -void intel_lvds_init(struct drm_i915_private *dev_priv);
> > -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv);
> > -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv);
> > +void intel_lvds_init(struct drm_i915_private *i915);
> > +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915);
> > +bool intel_is_dual_link_lvds(struct drm_i915_private *i915);
> >   #endif /* __INTEL_LVDS_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_opregion.c b/drivers/gpu/drm/i915/intel_opregion.c
> > index 8fa1159d097f..6cbe775d7537 100644
> > --- a/drivers/gpu/drm/i915/intel_opregion.c
> > +++ b/drivers/gpu/drm/i915/intel_opregion.c
> > @@ -266,11 +266,11 @@ struct opregion_asle_ext {
> >   #define MAX_DSLP	1500
> > -static int swsci(struct drm_i915_private *dev_priv,
> > +static int swsci(struct drm_i915_private *i915,
> >   		 u32 function, u32 parm, u32 *parm_out)
> >   {
> > -	struct opregion_swsci *swsci = dev_priv->opregion.swsci;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct opregion_swsci *swsci = i915->opregion.swsci;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u32 main_function, sub_function, scic;
> >   	u16 swsci_val;
> >   	u32 dslp;
> > @@ -285,11 +285,11 @@ static int swsci(struct drm_i915_private *dev_priv,
> >   	/* Check if we can call the function. See swsci_setup for details. */
> >   	if (main_function == SWSCI_SBCB) {
> > -		if ((dev_priv->opregion.swsci_sbcb_sub_functions &
> > +		if ((i915->opregion.swsci_sbcb_sub_functions &
> >   		     (1 << sub_function)) == 0)
> >   			return -EINVAL;
> >   	} else if (main_function == SWSCI_GBDA) {
> > -		if ((dev_priv->opregion.swsci_gbda_sub_functions &
> > +		if ((i915->opregion.swsci_gbda_sub_functions &
> >   		     (1 << sub_function)) == 0)
> >   			return -EINVAL;
> >   	}
> > @@ -363,13 +363,13 @@ static int swsci(struct drm_i915_private *dev_priv,
> >   int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
> >   				  bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> >   	u32 parm = 0;
> >   	u32 type = 0;
> >   	u32 port;
> >   	/* don't care about old stuff for now */
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		return 0;
> >   	if (intel_encoder->type == INTEL_OUTPUT_DSI)
> > @@ -409,7 +409,7 @@ int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
> >   	parm |= type << (16 + port * 3);
> > -	return swsci(dev_priv, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL);
> > +	return swsci(i915, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL);
> >   }
> >   static const struct {
> > @@ -423,29 +423,29 @@ static const struct {
> >   	{ PCI_D3cold,	0x04 },
> >   };
> > -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv,
> > +int intel_opregion_notify_adapter(struct drm_i915_private *i915,
> >   				  pci_power_t state)
> >   {
> >   	int i;
> > -	if (!HAS_DDI(dev_priv))
> > +	if (!HAS_DDI(i915))
> >   		return 0;
> >   	for (i = 0; i < ARRAY_SIZE(power_state_map); i++) {
> >   		if (state == power_state_map[i].pci_power_state)
> > -			return swsci(dev_priv, SWSCI_SBCB_ADAPTER_POWER_STATE,
> > +			return swsci(i915, SWSCI_SBCB_ADAPTER_POWER_STATE,
> >   				     power_state_map[i].parm, NULL);
> >   	}
> >   	return -EINVAL;
> >   }
> > -static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp)
> > +static u32 asle_set_backlight(struct drm_i915_private *i915, u32 bclp)
> >   {
> >   	struct intel_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> > -	struct opregion_asle *asle = dev_priv->opregion.asle;
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct opregion_asle *asle = i915->opregion.asle;
> > +	struct drm_device *dev = &i915->drm;
> >   	DRM_DEBUG_DRIVER("bclp = 0x%08x\n", bclp);
> > @@ -480,7 +480,7 @@ static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp)
> >   	return 0;
> >   }
> > -static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi)
> > +static u32 asle_set_als_illum(struct drm_i915_private *i915, u32 alsi)
> >   {
> >   	/* alsi is the current ALS reading in lux. 0 indicates below sensor
> >   	   range, 0xffff indicates above sensor range. 1-0xfffe are valid */
> > @@ -488,13 +488,13 @@ static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi)
> >   	return ASLC_ALS_ILLUM_FAILED;
> >   }
> > -static u32 asle_set_pwm_freq(struct drm_i915_private *dev_priv, u32 pfmb)
> > +static u32 asle_set_pwm_freq(struct drm_i915_private *i915, u32 pfmb)
> >   {
> >   	DRM_DEBUG_DRIVER("PWM freq is not supported\n");
> >   	return ASLC_PWM_FREQ_FAILED;
> >   }
> > -static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit)
> > +static u32 asle_set_pfit(struct drm_i915_private *i915, u32 pfit)
> >   {
> >   	/* Panel fitting is currently controlled by the X code, so this is a
> >   	   noop until modesetting support works fully */
> > @@ -502,13 +502,13 @@ static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit)
> >   	return ASLC_PFIT_FAILED;
> >   }
> > -static u32 asle_set_supported_rotation_angles(struct drm_i915_private *dev_priv, u32 srot)
> > +static u32 asle_set_supported_rotation_angles(struct drm_i915_private *i915, u32 srot)
> >   {
> >   	DRM_DEBUG_DRIVER("SROT is not supported\n");
> >   	return ASLC_ROTATION_ANGLES_FAILED;
> >   }
> > -static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer)
> > +static u32 asle_set_button_array(struct drm_i915_private *i915, u32 iuer)
> >   {
> >   	if (!iuer)
> >   		DRM_DEBUG_DRIVER("Button array event is not supported (nothing)\n");
> > @@ -526,7 +526,7 @@ static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer)
> >   	return ASLC_BUTTON_ARRAY_FAILED;
> >   }
> > -static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer)
> > +static u32 asle_set_convertible(struct drm_i915_private *i915, u32 iuer)
> >   {
> >   	if (iuer & ASLE_IUER_CONVERTIBLE)
> >   		DRM_DEBUG_DRIVER("Convertible is not supported (clamshell)\n");
> > @@ -536,7 +536,7 @@ static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer)
> >   	return ASLC_CONVERTIBLE_FAILED;
> >   }
> > -static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer)
> > +static u32 asle_set_docking(struct drm_i915_private *i915, u32 iuer)
> >   {
> >   	if (iuer & ASLE_IUER_DOCKING)
> >   		DRM_DEBUG_DRIVER("Docking is not supported (docked)\n");
> > @@ -546,7 +546,7 @@ static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer)
> >   	return ASLC_DOCKING_FAILED;
> >   }
> > -static u32 asle_isct_state(struct drm_i915_private *dev_priv)
> > +static u32 asle_isct_state(struct drm_i915_private *i915)
> >   {
> >   	DRM_DEBUG_DRIVER("ISCT is not supported\n");
> >   	return ASLC_ISCT_STATE_FAILED;
> > @@ -556,9 +556,9 @@ static void asle_work(struct work_struct *work)
> >   {
> >   	struct intel_opregion *opregion =
> >   		container_of(work, struct intel_opregion, asle_work);
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		container_of(opregion, struct drm_i915_private, opregion);
> > -	struct opregion_asle *asle = dev_priv->opregion.asle;
> > +	struct opregion_asle *asle = i915->opregion.asle;
> >   	u32 aslc_stat = 0;
> >   	u32 aslc_req;
> > @@ -574,40 +574,40 @@ static void asle_work(struct work_struct *work)
> >   	}
> >   	if (aslc_req & ASLC_SET_ALS_ILLUM)
> > -		aslc_stat |= asle_set_als_illum(dev_priv, asle->alsi);
> > +		aslc_stat |= asle_set_als_illum(i915, asle->alsi);
> >   	if (aslc_req & ASLC_SET_BACKLIGHT)
> > -		aslc_stat |= asle_set_backlight(dev_priv, asle->bclp);
> > +		aslc_stat |= asle_set_backlight(i915, asle->bclp);
> >   	if (aslc_req & ASLC_SET_PFIT)
> > -		aslc_stat |= asle_set_pfit(dev_priv, asle->pfit);
> > +		aslc_stat |= asle_set_pfit(i915, asle->pfit);
> >   	if (aslc_req & ASLC_SET_PWM_FREQ)
> > -		aslc_stat |= asle_set_pwm_freq(dev_priv, asle->pfmb);
> > +		aslc_stat |= asle_set_pwm_freq(i915, asle->pfmb);
> >   	if (aslc_req & ASLC_SUPPORTED_ROTATION_ANGLES)
> > -		aslc_stat |= asle_set_supported_rotation_angles(dev_priv,
> > +		aslc_stat |= asle_set_supported_rotation_angles(i915,
> >   							asle->srot);
> >   	if (aslc_req & ASLC_BUTTON_ARRAY)
> > -		aslc_stat |= asle_set_button_array(dev_priv, asle->iuer);
> > +		aslc_stat |= asle_set_button_array(i915, asle->iuer);
> >   	if (aslc_req & ASLC_CONVERTIBLE_INDICATOR)
> > -		aslc_stat |= asle_set_convertible(dev_priv, asle->iuer);
> > +		aslc_stat |= asle_set_convertible(i915, asle->iuer);
> >   	if (aslc_req & ASLC_DOCKING_INDICATOR)
> > -		aslc_stat |= asle_set_docking(dev_priv, asle->iuer);
> > +		aslc_stat |= asle_set_docking(i915, asle->iuer);
> >   	if (aslc_req & ASLC_ISCT_STATE_CHANGE)
> > -		aslc_stat |= asle_isct_state(dev_priv);
> > +		aslc_stat |= asle_isct_state(i915);
> >   	asle->aslc = aslc_stat;
> >   }
> > -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv)
> > +void intel_opregion_asle_intr(struct drm_i915_private *i915)
> >   {
> > -	if (dev_priv->opregion.asle)
> > -		schedule_work(&dev_priv->opregion.asle_work);
> > +	if (i915->opregion.asle)
> > +		schedule_work(&i915->opregion.asle_work);
> >   }
> >   #define ACPI_EV_DISPLAY_SWITCH (1<<0)
> > @@ -702,9 +702,9 @@ static u32 acpi_display_type(struct intel_connector *connector)
> >   	return display_type;
> >   }
> > -static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> > +static void intel_didl_outputs(struct drm_i915_private *i915)
> >   {
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > +	struct intel_opregion *opregion = &i915->opregion;
> >   	struct intel_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	int i = 0, max_outputs;
> > @@ -720,7 +720,7 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> >   	max_outputs = ARRAY_SIZE(opregion->acpi->didl) +
> >   		ARRAY_SIZE(opregion->acpi->did2);
> > -	drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter);
> > +	drm_connector_list_iter_begin(&i915->drm, &conn_iter);
> >   	for_each_intel_connector_iter(connector, &conn_iter) {
> >   		u32 device_id, type;
> > @@ -749,9 +749,9 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv)
> >   		set_did(opregion, i, 0);
> >   }
> > -static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> > +static void intel_setup_cadls(struct drm_i915_private *i915)
> >   {
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > +	struct intel_opregion *opregion = &i915->opregion;
> >   	struct intel_connector *connector;
> >   	struct drm_connector_list_iter conn_iter;
> >   	int i = 0;
> > @@ -766,7 +766,7 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> >   	 * Note that internal panels should be at the front of the connector
> >   	 * list already, ensuring they're not left out.
> >   	 */
> > -	drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter);
> > +	drm_connector_list_iter_begin(&i915->drm, &conn_iter);
> >   	for_each_intel_connector_iter(connector, &conn_iter) {
> >   		if (i >= ARRAY_SIZE(opregion->acpi->cadl))
> >   			break;
> > @@ -779,9 +779,9 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv)
> >   		opregion->acpi->cadl[i] = 0;
> >   }
> > -static void swsci_setup(struct drm_i915_private *dev_priv)
> > +static void swsci_setup(struct drm_i915_private *i915)
> >   {
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > +	struct intel_opregion *opregion = &i915->opregion;
> >   	bool requested_callbacks = false;
> >   	u32 tmp;
> > @@ -790,7 +790,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> >   	opregion->swsci_sbcb_sub_functions = 1;
> >   	/* We use GBDA to ask for supported GBDA calls. */
> > -	if (swsci(dev_priv, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) {
> > +	if (swsci(i915, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) {
> >   		/* make the bits match the sub-function codes */
> >   		tmp <<= 1;
> >   		opregion->swsci_gbda_sub_functions |= tmp;
> > @@ -801,7 +801,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> >   	 * must not call interfaces that are not specifically requested by the
> >   	 * bios.
> >   	 */
> > -	if (swsci(dev_priv, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) {
> > +	if (swsci(i915, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) {
> >   		/* here, the bits already match sub-function codes */
> >   		opregion->swsci_sbcb_sub_functions |= tmp;
> >   		requested_callbacks = true;
> > @@ -812,7 +812,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv)
> >   	 * the callback is _requested_. But we still can't call interfaces that
> >   	 * are not requested.
> >   	 */
> > -	if (swsci(dev_priv, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) {
> > +	if (swsci(i915, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) {
> >   		/* make the bits match the sub-function codes */
> >   		u32 low = tmp & 0x7ff;
> >   		u32 high = tmp & ~0xfff; /* bit 11 is reserved */
> > @@ -854,9 +854,9 @@ static const struct dmi_system_id intel_no_opregion_vbt[] = {
> >   	{ }
> >   };
> > -static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> > +static int intel_load_vbt_firmware(struct drm_i915_private *i915)
> >   {
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > +	struct intel_opregion *opregion = &i915->opregion;
> >   	const struct firmware *fw = NULL;
> >   	const char *name = i915_modparams.vbt_firmware;
> >   	int ret;
> > @@ -864,7 +864,7 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> >   	if (!name || !*name)
> >   		return -ENOENT;
> > -	ret = request_firmware(&fw, name, &dev_priv->drm.pdev->dev);
> > +	ret = request_firmware(&fw, name, &i915->drm.pdev->dev);
> >   	if (ret) {
> >   		DRM_ERROR("Requesting VBT firmware \"%s\" failed (%d)\n",
> >   			  name, ret);
> > @@ -891,10 +891,10 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv)
> >   	return ret;
> >   }
> > -int intel_opregion_setup(struct drm_i915_private *dev_priv)
> > +int intel_opregion_setup(struct drm_i915_private *i915)
> >   {
> > -	struct intel_opregion *opregion = &dev_priv->opregion;
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct intel_opregion *opregion = &i915->opregion;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u32 asls, mboxes;
> >   	char buf[sizeof(OPREGION_SIGNATURE)];
> >   	int err = 0;
> > @@ -945,7 +945,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv)
> >   	if (mboxes & MBOX_SWSCI) {
> >   		DRM_DEBUG_DRIVER("SWSCI supported\n");
> >   		opregion->swsci = base + OPREGION_SWSCI_OFFSET;
> > -		swsci_setup(dev_priv);
> > +		swsci_setup(i915);
> >   	}
> >   	if (mboxes & MBOX_ASLE) {
> > @@ -958,7 +958,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv)
> >   	if (mboxes & MBOX_ASLE_EXT)
> >   		DRM_DEBUG_DRIVER("ASLE extension supported\n");
> > -	if (intel_load_vbt_firmware(dev_priv) == 0)
> > +	if (intel_load_vbt_firmware(i915) == 0)
> >   		goto out;
> >   	if (dmi_check_system(intel_no_opregion_vbt))
> > @@ -1043,12 +1043,12 @@ static const struct dmi_system_id intel_use_opregion_panel_type[] = {
> >   };
> >   int
> > -intel_opregion_get_panel_type(struct drm_i915_private *dev_priv)
> > +intel_opregion_get_panel_type(struct drm_i915_private *i915)
> >   {
> >   	u32 panel_details;
> >   	int ret;
> > -	ret = swsci(dev_priv, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details);
> > +	ret = swsci(i915, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details);
> >   	if (ret) {
> >   		DRM_DEBUG_KMS("Failed to get panel details from OpRegion (%d)\n",
> >   			      ret);
> > diff --git a/drivers/gpu/drm/i915/intel_opregion.h b/drivers/gpu/drm/i915/intel_opregion.h
> > index 4aa68ffbd30e..8e0b7bdecbe4 100644
> > --- a/drivers/gpu/drm/i915/intel_opregion.h
> > +++ b/drivers/gpu/drm/i915/intel_opregion.h
> > @@ -56,47 +56,47 @@ struct intel_opregion {
> >   #ifdef CONFIG_ACPI
> > -int intel_opregion_setup(struct drm_i915_private *dev_priv);
> > +int intel_opregion_setup(struct drm_i915_private *i915);
> > -void intel_opregion_register(struct drm_i915_private *dev_priv);
> > -void intel_opregion_unregister(struct drm_i915_private *dev_priv);
> > +void intel_opregion_register(struct drm_i915_private *i915);
> > +void intel_opregion_unregister(struct drm_i915_private *i915);
> > -void intel_opregion_resume(struct drm_i915_private *dev_priv);
> > -void intel_opregion_suspend(struct drm_i915_private *dev_priv,
> > +void intel_opregion_resume(struct drm_i915_private *i915);
> > +void intel_opregion_suspend(struct drm_i915_private *i915,
> >   			    pci_power_t state);
> > -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv);
> > +void intel_opregion_asle_intr(struct drm_i915_private *i915);
> >   int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder,
> >   				  bool enable);
> > -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv,
> > +int intel_opregion_notify_adapter(struct drm_i915_private *i915,
> >   				  pci_power_t state);
> > -int intel_opregion_get_panel_type(struct drm_i915_private *dev_priv);
> > +int intel_opregion_get_panel_type(struct drm_i915_private *i915);
> >   #else /* CONFIG_ACPI*/
> > -static inline int intel_opregion_setup(struct drm_i915_private *dev_priv)
> > +static inline int intel_opregion_setup(struct drm_i915_private *i915)
> >   {
> >   	return 0;
> >   }
> > -static inline void intel_opregion_register(struct drm_i915_private *dev_priv)
> > +static inline void intel_opregion_register(struct drm_i915_private *i915)
> >   {
> >   }
> > -static inline void intel_opregion_unregister(struct drm_i915_private *dev_priv)
> > +static inline void intel_opregion_unregister(struct drm_i915_private *i915)
> >   {
> >   }
> > -static inline void intel_opregion_resume(struct drm_i915_private *dev_priv)
> > +static inline void intel_opregion_resume(struct drm_i915_private *i915)
> >   {
> >   }
> > -static inline void intel_opregion_suspend(struct drm_i915_private *dev_priv,
> > +static inline void intel_opregion_suspend(struct drm_i915_private *i915,
> >   					  pci_power_t state)
> >   {
> >   }
> > -static inline void intel_opregion_asle_intr(struct drm_i915_private *dev_priv)
> > +static inline void intel_opregion_asle_intr(struct drm_i915_private *i915)
> >   {
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c
> > index a2ac06a08715..802bc4c72ecd 100644
> > --- a/drivers/gpu/drm/i915/intel_overlay.c
> > +++ b/drivers/gpu/drm/i915/intel_overlay.c
> > @@ -193,10 +193,10 @@ struct intel_overlay {
> >   	struct i915_active_request last_flip;
> >   };
> > -static void i830_overlay_clock_gating(struct drm_i915_private *dev_priv,
> > +static void i830_overlay_clock_gating(struct drm_i915_private *i915,
> >   				      bool enable)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	u8 val;
> >   	/* WA_OVERLAY_CLKGATE:alm */
> > @@ -247,7 +247,7 @@ static struct i915_request *alloc_request(struct intel_overlay *overlay)
> >   /* overlay needs to be disable in OCMD reg */
> >   static int intel_overlay_on(struct intel_overlay *overlay)
> >   {
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	struct i915_request *rq;
> >   	u32 *cs;
> > @@ -265,8 +265,8 @@ static int intel_overlay_on(struct intel_overlay *overlay)
> >   	overlay->active = true;
> > -	if (IS_I830(dev_priv))
> > -		i830_overlay_clock_gating(dev_priv, false);
> > +	if (IS_I830(i915))
> > +		i830_overlay_clock_gating(i915, false);
> >   	*cs++ = MI_OVERLAY_FLIP | MI_OVERLAY_ON;
> >   	*cs++ = overlay->flip_addr | OFC_UPDATE;
> > @@ -303,7 +303,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay,
> >   				  struct i915_vma *vma,
> >   				  bool load_polyphase_filter)
> >   {
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	struct i915_request *rq;
> >   	u32 flip_addr = overlay->flip_addr;
> >   	u32 tmp, *cs;
> > @@ -369,7 +369,7 @@ static void intel_overlay_off_tail(struct i915_active_request *active,
> >   {
> >   	struct intel_overlay *overlay =
> >   		container_of(active, typeof(*overlay), last_flip);
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	intel_overlay_release_old_vma(overlay);
> > @@ -377,8 +377,8 @@ static void intel_overlay_off_tail(struct i915_active_request *active,
> >   	overlay->crtc = NULL;
> >   	overlay->active = false;
> > -	if (IS_I830(dev_priv))
> > -		i830_overlay_clock_gating(dev_priv, true);
> > +	if (IS_I830(i915))
> > +		i830_overlay_clock_gating(i915, true);
> >   }
> >   /* overlay needs to be disabled in OCMD reg */
> > @@ -437,11 +437,11 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay)
> >    */
> >   static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
> >   {
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	u32 *cs;
> >   	int ret;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> >   	/* Only wait if there is actually an old frame to release to
> >   	 * guarantee forward progress.
> > @@ -477,9 +477,9 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
> >   	return 0;
> >   }
> > -void intel_overlay_reset(struct drm_i915_private *dev_priv)
> > +void intel_overlay_reset(struct drm_i915_private *i915)
> >   {
> > -	struct intel_overlay *overlay = dev_priv->overlay;
> > +	struct intel_overlay *overlay = i915->overlay;
> >   	if (!overlay)
> >   		return;
> > @@ -540,11 +540,11 @@ static int uv_vsubsampling(u32 format)
> >   	}
> >   }
> > -static u32 calc_swidthsw(struct drm_i915_private *dev_priv, u32 offset, u32 width)
> > +static u32 calc_swidthsw(struct drm_i915_private *i915, u32 offset, u32 width)
> >   {
> >   	u32 sw;
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		sw = ALIGN((offset & 31) + width, 32);
> >   	else
> >   		sw = ALIGN((offset & 63) + width, 64);
> > @@ -749,21 +749,21 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> >   				      struct drm_intel_overlay_put_image *params)
> >   {
> >   	struct overlay_registers __iomem *regs = overlay->regs;
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	u32 swidth, swidthsw, sheight, ostride;
> >   	enum pipe pipe = overlay->crtc->pipe;
> >   	bool scale_changed = false;
> >   	struct i915_vma *vma;
> >   	int ret, tmp_width;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > -	WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> > +	WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
> >   	ret = intel_overlay_release_old_vid(overlay);
> >   	if (ret != 0)
> >   		return ret;
> > -	atomic_inc(&dev_priv->gpu_error.pending_fb_pin);
> > +	atomic_inc(&i915->gpu_error.pending_fb_pin);
> >   	i915_gem_object_lock(new_bo);
> >   	vma = i915_gem_object_pin_to_display_plane(new_bo,
> > @@ -783,7 +783,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> >   		u32 oconfig;
> >   		oconfig = OCONF_CC_OUT_8BIT;
> > -		if (IS_GEN(dev_priv, 4))
> > +		if (IS_GEN(i915, 4))
> >   			oconfig |= OCONF_CSC_MODE_BT709;
> >   		oconfig |= pipe == 0 ?
> >   			OCONF_PIPE_A : OCONF_PIPE_B;
> > @@ -804,7 +804,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> >   		tmp_width = params->src_width;
> >   	swidth = params->src_width;
> > -	swidthsw = calc_swidthsw(dev_priv, params->offset_Y, tmp_width);
> > +	swidthsw = calc_swidthsw(i915, params->offset_Y, tmp_width);
> >   	sheight = params->src_height;
> >   	iowrite32(i915_ggtt_offset(vma) + params->offset_Y, &regs->OBUF_0Y);
> >   	ostride = params->stride_Y;
> > @@ -817,9 +817,9 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> >   		swidth |= (params->src_width / uv_hscale) << 16;
> >   		sheight |= (params->src_height / uv_vscale) << 16;
> > -		tmp_U = calc_swidthsw(dev_priv, params->offset_U,
> > +		tmp_U = calc_swidthsw(i915, params->offset_U,
> >   				      params->src_width / uv_hscale);
> > -		tmp_V = calc_swidthsw(dev_priv, params->offset_V,
> > +		tmp_V = calc_swidthsw(i915, params->offset_V,
> >   				      params->src_width / uv_hscale);
> >   		swidthsw |= max(tmp_U, tmp_V) << 16;
> > @@ -851,18 +851,18 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay,
> >   out_unpin:
> >   	i915_gem_object_unpin_from_display_plane(vma);
> >   out_pin_section:
> > -	atomic_dec(&dev_priv->gpu_error.pending_fb_pin);
> > +	atomic_dec(&i915->gpu_error.pending_fb_pin);
> >   	return ret;
> >   }
> >   int intel_overlay_switch_off(struct intel_overlay *overlay)
> >   {
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	int ret;
> > -	lockdep_assert_held(&dev_priv->drm.struct_mutex);
> > -	WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex));
> > +	lockdep_assert_held(&i915->drm.struct_mutex);
> > +	WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex));
> >   	ret = intel_overlay_recover_from_interrupt(overlay);
> >   	if (ret != 0)
> > @@ -895,14 +895,14 @@ static int check_overlay_possible_on_crtc(struct intel_overlay *overlay,
> >   static void update_pfit_vscale_ratio(struct intel_overlay *overlay)
> >   {
> > -	struct drm_i915_private *dev_priv = overlay->i915;
> > +	struct drm_i915_private *i915 = overlay->i915;
> >   	u32 pfit_control = I915_READ(PFIT_CONTROL);
> >   	u32 ratio;
> >   	/* XXX: This is not the same logic as in the xorg driver, but more in
> >   	 * line with the intel documentation for the i965
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		/* on i965 use the PGM reg to read out the autoscaler values */
> >   		ratio = I915_READ(PFIT_PGM_RATIOS) >> PFIT_VERT_SCALE_SHIFT_965;
> >   	} else {
> > @@ -947,7 +947,7 @@ static int check_overlay_scaling(struct drm_intel_overlay_put_image *rec)
> >   	return 0;
> >   }
> > -static int check_overlay_src(struct drm_i915_private *dev_priv,
> > +static int check_overlay_src(struct drm_i915_private *i915,
> >   			     struct drm_intel_overlay_put_image *rec,
> >   			     struct drm_i915_gem_object *new_bo)
> >   {
> > @@ -958,7 +958,7 @@ static int check_overlay_src(struct drm_i915_private *dev_priv,
> >   	u32 tmp;
> >   	/* check src dimensions */
> > -	if (IS_I845G(dev_priv) || IS_I830(dev_priv)) {
> > +	if (IS_I845G(i915) || IS_I830(i915)) {
> >   		if (rec->src_height > IMAGE_MAX_HEIGHT_LEGACY ||
> >   		    rec->src_width  > IMAGE_MAX_WIDTH_LEGACY)
> >   			return -EINVAL;
> > @@ -1010,14 +1010,14 @@ static int check_overlay_src(struct drm_i915_private *dev_priv,
> >   		return -EINVAL;
> >   	/* stride checking */
> > -	if (IS_I830(dev_priv) || IS_I845G(dev_priv))
> > +	if (IS_I830(i915) || IS_I845G(i915))
> >   		stride_mask = 255;
> >   	else
> >   		stride_mask = 63;
> >   	if (rec->stride_Y & stride_mask || rec->stride_UV & stride_mask)
> >   		return -EINVAL;
> > -	if (IS_GEN(dev_priv, 4) && rec->stride_Y < 512)
> > +	if (IS_GEN(i915, 4) && rec->stride_Y < 512)
> >   		return -EINVAL;
> >   	tmp = (rec->flags & I915_OVERLAY_TYPE_MASK) == I915_OVERLAY_YUV_PLANAR ?
> > @@ -1062,14 +1062,14 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> >   				  struct drm_file *file_priv)
> >   {
> >   	struct drm_intel_overlay_put_image *params = data;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_overlay *overlay;
> >   	struct drm_crtc *drmmode_crtc;
> >   	struct intel_crtc *crtc;
> >   	struct drm_i915_gem_object *new_bo;
> >   	int ret;
> > -	overlay = dev_priv->overlay;
> > +	overlay = i915->overlay;
> >   	if (!overlay) {
> >   		DRM_DEBUG("userspace bug: no overlay\n");
> >   		return -ENODEV;
> > @@ -1148,7 +1148,7 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> >   		goto out_unlock;
> >   	}
> > -	ret = check_overlay_src(dev_priv, params, new_bo);
> > +	ret = check_overlay_src(i915, params, new_bo);
> >   	if (ret != 0)
> >   		goto out_unlock;
> > @@ -1231,11 +1231,11 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> >   			      struct drm_file *file_priv)
> >   {
> >   	struct drm_intel_overlay_attrs *attrs = data;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_overlay *overlay;
> >   	int ret;
> > -	overlay = dev_priv->overlay;
> > +	overlay = i915->overlay;
> >   	if (!overlay) {
> >   		DRM_DEBUG("userspace bug: no overlay\n");
> >   		return -ENODEV;
> > @@ -1251,7 +1251,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> >   		attrs->contrast   = overlay->contrast;
> >   		attrs->saturation = overlay->saturation;
> > -		if (!IS_GEN(dev_priv, 2)) {
> > +		if (!IS_GEN(i915, 2)) {
> >   			attrs->gamma0 = I915_READ(OGAMC0);
> >   			attrs->gamma1 = I915_READ(OGAMC1);
> >   			attrs->gamma2 = I915_READ(OGAMC2);
> > @@ -1275,7 +1275,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> >   		update_reg_attrs(overlay, overlay->regs);
> >   		if (attrs->flags & I915_OVERLAY_UPDATE_GAMMA) {
> > -			if (IS_GEN(dev_priv, 2))
> > +			if (IS_GEN(i915, 2))
> >   				goto out_unlock;
> >   			if (overlay->active) {
> > @@ -1351,19 +1351,19 @@ static int get_registers(struct intel_overlay *overlay, bool use_phys)
> >   	return err;
> >   }
> > -void intel_overlay_setup(struct drm_i915_private *dev_priv)
> > +void intel_overlay_setup(struct drm_i915_private *i915)
> >   {
> >   	struct intel_overlay *overlay;
> >   	int ret;
> > -	if (!HAS_OVERLAY(dev_priv))
> > +	if (!HAS_OVERLAY(i915))
> >   		return;
> >   	overlay = kzalloc(sizeof(*overlay), GFP_KERNEL);
> >   	if (!overlay)
> >   		return;
> > -	overlay->i915 = dev_priv;
> > +	overlay->i915 = i915;
> >   	overlay->color_key = 0x0101fe;
> >   	overlay->color_key_enabled = true;
> > @@ -1373,7 +1373,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
> >   	INIT_ACTIVE_REQUEST(&overlay->last_flip);
> > -	ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(dev_priv));
> > +	ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(i915));
> >   	if (ret)
> >   		goto out_free;
> > @@ -1387,7 +1387,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
> >   	update_polyphase_filter(overlay->regs);
> >   	update_reg_attrs(overlay, overlay->regs);
> > -	dev_priv->overlay = overlay;
> > +	i915->overlay = overlay;
> >   	DRM_INFO("Initialized overlay support.\n");
> >   	return;
> > @@ -1397,11 +1397,11 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv)
> >   	kfree(overlay);
> >   }
> > -void intel_overlay_cleanup(struct drm_i915_private *dev_priv)
> > +void intel_overlay_cleanup(struct drm_i915_private *i915)
> >   {
> >   	struct intel_overlay *overlay;
> > -	overlay = fetch_and_zero(&dev_priv->overlay);
> > +	overlay = fetch_and_zero(&i915->overlay);
> >   	if (!overlay)
> >   		return;
> > @@ -1427,9 +1427,9 @@ struct intel_overlay_error_state {
> >   };
> >   struct intel_overlay_error_state *
> > -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv)
> > +intel_overlay_capture_error_state(struct drm_i915_private *i915)
> >   {
> > -	struct intel_overlay *overlay = dev_priv->overlay;
> > +	struct intel_overlay *overlay = i915->overlay;
> >   	struct intel_overlay_error_state *error;
> >   	if (!overlay || !overlay->active)
> > diff --git a/drivers/gpu/drm/i915/intel_overlay.h b/drivers/gpu/drm/i915/intel_overlay.h
> > index a167c28acd27..2e8771cffee9 100644
> > --- a/drivers/gpu/drm/i915/intel_overlay.h
> > +++ b/drivers/gpu/drm/i915/intel_overlay.h
> > @@ -13,16 +13,16 @@ struct drm_i915_private;
> >   struct intel_overlay;
> >   struct intel_overlay_error_state;
> > -void intel_overlay_setup(struct drm_i915_private *dev_priv);
> > -void intel_overlay_cleanup(struct drm_i915_private *dev_priv);
> > +void intel_overlay_setup(struct drm_i915_private *i915);
> > +void intel_overlay_cleanup(struct drm_i915_private *i915);
> >   int intel_overlay_switch_off(struct intel_overlay *overlay);
> >   int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data,
> >   				  struct drm_file *file_priv);
> >   int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data,
> >   			      struct drm_file *file_priv);
> > -void intel_overlay_reset(struct drm_i915_private *dev_priv);
> > +void intel_overlay_reset(struct drm_i915_private *i915);
> >   struct intel_overlay_error_state *
> > -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv);
> > +intel_overlay_capture_error_state(struct drm_i915_private *i915);
> >   void intel_overlay_print_error_state(struct drm_i915_error_state_buf *e,
> >   				     struct intel_overlay_error_state *error);
> > diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c
> > index 9cd4e37e3934..d80115089445 100644
> > --- a/drivers/gpu/drm/i915/intel_panel.c
> > +++ b/drivers/gpu/drm/i915/intel_panel.c
> > @@ -65,7 +65,7 @@ struct drm_display_mode *
> >   intel_panel_edid_downclock_mode(struct intel_connector *connector,
> >   				const struct drm_display_mode *fixed_mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	const struct drm_display_mode *scan, *best_mode = NULL;
> >   	struct drm_display_mode *downclock_mode;
> >   	int best_clock = fixed_mode->clock;
> > @@ -92,7 +92,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector,
> >   	if (!best_mode)
> >   		return NULL;
> > -	downclock_mode = drm_mode_duplicate(&dev_priv->drm, best_mode);
> > +	downclock_mode = drm_mode_duplicate(&i915->drm, best_mode);
> >   	if (!downclock_mode)
> >   		return NULL;
> > @@ -106,7 +106,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector,
> >   struct drm_display_mode *
> >   intel_panel_edid_fixed_mode(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	const struct drm_display_mode *scan;
> >   	struct drm_display_mode *fixed_mode;
> > @@ -118,7 +118,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> >   		if ((scan->type & DRM_MODE_TYPE_PREFERRED) == 0)
> >   			continue;
> > -		fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan);
> > +		fixed_mode = drm_mode_duplicate(&i915->drm, scan);
> >   		if (!fixed_mode)
> >   			return NULL;
> > @@ -132,7 +132,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> >   	scan = list_first_entry(&connector->base.probed_modes,
> >   				typeof(*scan), head);
> > -	fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan);
> > +	fixed_mode = drm_mode_duplicate(&i915->drm, scan);
> >   	if (!fixed_mode)
> >   		return NULL;
> > @@ -148,15 +148,15 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector)
> >   struct drm_display_mode *
> >   intel_panel_vbt_fixed_mode(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct drm_display_info *info = &connector->base.display_info;
> >   	struct drm_display_mode *fixed_mode;
> > -	if (!dev_priv->vbt.lfp_lvds_vbt_mode)
> > +	if (!i915->vbt.lfp_lvds_vbt_mode)
> >   		return NULL;
> > -	fixed_mode = drm_mode_duplicate(&dev_priv->drm,
> > -					dev_priv->vbt.lfp_lvds_vbt_mode);
> > +	fixed_mode = drm_mode_duplicate(&i915->drm,
> > +					i915->vbt.lfp_lvds_vbt_mode);
> >   	if (!fixed_mode)
> >   		return NULL;
> > @@ -378,7 +378,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> >   			      struct intel_crtc_state *pipe_config,
> >   			      int fitting_mode)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	u32 pfit_control = 0, pfit_pgm_ratios = 0, border = 0;
> >   	struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> > @@ -399,7 +399,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> >   		break;
> >   	case DRM_MODE_SCALE_ASPECT:
> >   		/* Scale but preserve the aspect ratio */
> > -		if (INTEL_GEN(dev_priv) >= 4)
> > +		if (INTEL_GEN(i915) >= 4)
> >   			i965_scale_aspect(pipe_config, &pfit_control);
> >   		else
> >   			i9xx_scale_aspect(pipe_config, &pfit_control,
> > @@ -413,7 +413,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> >   		if (pipe_config->pipe_src_h != adjusted_mode->crtc_vdisplay ||
> >   		    pipe_config->pipe_src_w != adjusted_mode->crtc_hdisplay) {
> >   			pfit_control |= PFIT_ENABLE;
> > -			if (INTEL_GEN(dev_priv) >= 4)
> > +			if (INTEL_GEN(i915) >= 4)
> >   				pfit_control |= PFIT_SCALING_AUTO;
> >   			else
> >   				pfit_control |= (VERT_AUTO_SCALE |
> > @@ -429,7 +429,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> >   	/* 965+ wants fuzzy fitting */
> >   	/* FIXME: handle multiple panels by failing gracefully */
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		pfit_control |= ((intel_crtc->pipe << PFIT_PIPE_SHIFT) |
> >   				 PFIT_FILTER_FUZZY);
> > @@ -440,7 +440,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc,
> >   	}
> >   	/* Make sure pre-965 set dither correctly for 18bpp panels. */
> > -	if (INTEL_GEN(dev_priv) < 4 && pipe_config->pipe_bpp == 18)
> > +	if (INTEL_GEN(i915) < 4 && pipe_config->pipe_bpp == 18)
> >   		pfit_control |= PANEL_8TO6_DITHER_ENABLE;
> >   	pipe_config->gmch_pfit.control = pfit_control;
> > @@ -517,7 +517,7 @@ static inline u32 scale_hw_to_user(struct intel_connector *connector,
> >   static u32 intel_panel_compute_brightness(struct intel_connector *connector,
> >   					  u32 val)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	WARN_ON(panel->backlight.max == 0);
> > @@ -526,7 +526,7 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector,
> >   		return val;
> >   	if (i915_modparams.invert_brightness > 0 ||
> > -	    dev_priv->quirks & QUIRK_INVERT_BRIGHTNESS) {
> > +	    i915->quirks & QUIRK_INVERT_BRIGHTNESS) {
> >   		return panel->backlight.max - val + panel->backlight.min;
> >   	}
> > @@ -535,39 +535,39 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector,
> >   static u32 lpt_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	return I915_READ(BLC_PWM_PCH_CTL2) & BACKLIGHT_DUTY_CYCLE_MASK;
> >   }
> >   static u32 pch_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	return I915_READ(BLC_PWM_CPU_CTL) & BACKLIGHT_DUTY_CYCLE_MASK;
> >   }
> >   static u32 i9xx_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 val;
> >   	val = I915_READ(BLC_PWM_CTL) & BACKLIGHT_DUTY_CYCLE_MASK;
> > -	if (INTEL_GEN(dev_priv) < 4)
> > +	if (INTEL_GEN(i915) < 4)
> >   		val >>= 1;
> >   	if (panel->backlight.combination_mode) {
> >   		u8 lbpc;
> > -		pci_read_config_byte(dev_priv->drm.pdev, LBPC, &lbpc);
> > +		pci_read_config_byte(i915->drm.pdev, LBPC, &lbpc);
> >   		val *= lbpc;
> >   	}
> >   	return val;
> >   }
> > -static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe)
> > +static u32 _vlv_get_backlight(struct drm_i915_private *i915, enum pipe pipe)
> >   {
> >   	if (WARN_ON(pipe != PIPE_A && pipe != PIPE_B))
> >   		return 0;
> > @@ -577,15 +577,15 @@ static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe)
> >   static u32 vlv_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum pipe pipe = intel_connector_get_pipe(connector);
> > -	return _vlv_get_backlight(dev_priv, pipe);
> > +	return _vlv_get_backlight(i915, pipe);
> >   }
> >   static u32 bxt_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	return I915_READ(BXT_BLC_PWM_DUTY(panel->backlight.controller));
> > @@ -603,7 +603,7 @@ static u32 pwm_get_backlight(struct intel_connector *connector)
> >   static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	u32 val = I915_READ(BLC_PWM_PCH_CTL2) & ~BACKLIGHT_DUTY_CYCLE_MASK;
> >   	I915_WRITE(BLC_PWM_PCH_CTL2, val | level);
> > @@ -612,7 +612,7 @@ static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32
> >   static void pch_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	u32 tmp;
> >   	tmp = I915_READ(BLC_PWM_CPU_CTL) & ~BACKLIGHT_DUTY_CYCLE_MASK;
> > @@ -622,7 +622,7 @@ static void pch_set_backlight(const struct drm_connector_state *conn_state, u32
> >   static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 tmp, mask;
> > @@ -633,10 +633,10 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32
> >   		lbpc = level * 0xfe / panel->backlight.max + 1;
> >   		level /= lbpc;
> > -		pci_write_config_byte(dev_priv->drm.pdev, LBPC, lbpc);
> > +		pci_write_config_byte(i915->drm.pdev, LBPC, lbpc);
> >   	}
> > -	if (IS_GEN(dev_priv, 4)) {
> > +	if (IS_GEN(i915, 4)) {
> >   		mask = BACKLIGHT_DUTY_CYCLE_MASK;
> >   	} else {
> >   		level <<= 1;
> > @@ -650,7 +650,7 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32
> >   static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe;
> >   	u32 tmp;
> > @@ -661,7 +661,7 @@ static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32
> >   static void bxt_set_backlight(const struct drm_connector_state *conn_state, u32 level)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	I915_WRITE(BXT_BLC_PWM_DUTY(panel->backlight.controller), level);
> > @@ -694,7 +694,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> >   				    u32 user_level, u32 user_max)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 hw_level;
> > @@ -707,7 +707,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> >   	if (!panel->backlight.present || !conn_state->crtc)
> >   		return;
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	WARN_ON(panel->backlight.max == 0);
> > @@ -723,13 +723,13 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state
> >   	if (panel->backlight.enabled)
> >   		intel_panel_actually_set_backlight(conn_state, hw_level);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   }
> >   static void lpt_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	u32 tmp;
> >   	intel_panel_actually_set_backlight(old_conn_state, 0);
> > @@ -755,7 +755,7 @@ static void lpt_disable_backlight(const struct drm_connector_state *old_conn_sta
> >   static void pch_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	u32 tmp;
> >   	intel_panel_actually_set_backlight(old_conn_state, 0);
> > @@ -774,7 +774,7 @@ static void i9xx_disable_backlight(const struct drm_connector_state *old_conn_st
> >   static void i965_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(old_conn_state->connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(old_conn_state->connector->dev);
> >   	u32 tmp;
> >   	intel_panel_actually_set_backlight(old_conn_state, 0);
> > @@ -786,7 +786,7 @@ static void i965_disable_backlight(const struct drm_connector_state *old_conn_st
> >   static void vlv_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum pipe pipe = to_intel_crtc(old_conn_state->crtc)->pipe;
> >   	u32 tmp;
> > @@ -799,7 +799,7 @@ static void vlv_disable_backlight(const struct drm_connector_state *old_conn_sta
> >   static void bxt_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 tmp, val;
> > @@ -819,7 +819,7 @@ static void bxt_disable_backlight(const struct drm_connector_state *old_conn_sta
> >   static void cnp_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 tmp;
> > @@ -844,7 +844,7 @@ static void pwm_disable_backlight(const struct drm_connector_state *old_conn_sta
> >   void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(old_conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	if (!panel->backlight.present)
> > @@ -856,26 +856,26 @@ void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_st
> >   	 * backlight. This will leave the backlight on unnecessarily when
> >   	 * another client is not activated.
> >   	 */
> > -	if (dev_priv->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) {
> > +	if (i915->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) {
> >   		DRM_DEBUG_DRIVER("Skipping backlight disable on vga switch\n");
> >   		return;
> >   	}
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	if (panel->backlight.device)
> >   		panel->backlight.device->props.power = FB_BLANK_POWERDOWN;
> >   	panel->backlight.enabled = false;
> >   	panel->backlight.disable(old_conn_state);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   }
> >   static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 pch_ctl1, pch_ctl2, schicken;
> > @@ -886,7 +886,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   		I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1);
> >   	}
> > -	if (HAS_PCH_LPT(dev_priv)) {
> > +	if (HAS_PCH_LPT(i915)) {
> >   		schicken = I915_READ(SOUTH_CHICKEN2);
> >   		if (panel->backlight.alternate_pwm_increment)
> >   			schicken |= LPT_PWM_GRANULARITY;
> > @@ -910,7 +910,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   		pch_ctl1 |= BLM_PCH_POLARITY;
> >   	/* After LPT, override is the default. */
> > -	if (HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_LPT(i915))
> >   		pch_ctl1 |= BLM_PCH_OVERRIDE_ENABLE;
> >   	I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1);
> > @@ -925,7 +925,7 @@ static void pch_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 cpu_ctl2, pch_ctl1, pch_ctl2;
> > @@ -971,7 +971,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 ctl, freq;
> > @@ -988,7 +988,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   	ctl = freq << 17;
> >   	if (panel->backlight.combination_mode)
> >   		ctl |= BLM_LEGACY_MODE;
> > -	if (IS_PINEVIEW(dev_priv) && panel->backlight.active_low_pwm)
> > +	if (IS_PINEVIEW(i915) && panel->backlight.active_low_pwm)
> >   		ctl |= BLM_POLARITY_PNV;
> >   	I915_WRITE(BLC_PWM_CTL, ctl);
> > @@ -1002,7 +1002,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   	 * 855gm only, but checking for gen2 is safe, as 855gm is the only gen2
> >   	 * that has backlight.
> >   	 */
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		I915_WRITE(BLC_HIST_CTL, BLM_HISTOGRAM_ENABLE);
> >   }
> > @@ -1010,7 +1010,7 @@ static void i965_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe;
> >   	u32 ctl, ctl2, freq;
> > @@ -1045,7 +1045,7 @@ static void vlv_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> >   	u32 ctl, ctl2;
> > @@ -1075,7 +1075,7 @@ static void bxt_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> >   	u32 pwm_ctl, val;
> > @@ -1123,7 +1123,7 @@ static void cnp_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				 const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 pwm_ctl;
> > @@ -1187,7 +1187,7 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> > @@ -1196,28 +1196,28 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state,
> >   	DRM_DEBUG_KMS("pipe %c\n", pipe_name(pipe));
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	__intel_panel_enable_backlight(crtc_state, conn_state);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   }
> >   #if IS_ENABLED(CONFIG_BACKLIGHT_CLASS_DEVICE)
> >   static u32 intel_panel_get_backlight(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 val = 0;
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	if (panel->backlight.enabled) {
> >   		val = panel->backlight.get(connector);
> >   		val = intel_panel_compute_brightness(connector, val);
> >   	}
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   	DRM_DEBUG_DRIVER("get backlight PWM = %d\n", val);
> >   	return val;
> > @@ -1228,14 +1228,14 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta
> >   				      u32 user_level, u32 user_max)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 hw_level;
> >   	if (!panel->backlight.present)
> >   		return;
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	WARN_ON(panel->backlight.max == 0);
> > @@ -1245,7 +1245,7 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta
> >   	if (panel->backlight.enabled)
> >   		intel_panel_actually_set_backlight(conn_state, hw_level);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   }
> >   static int intel_backlight_device_update_status(struct backlight_device *bd)
> > @@ -1284,11 +1284,11 @@ static int intel_backlight_device_get_brightness(struct backlight_device *bd)
> >   {
> >   	struct intel_connector *connector = bl_get_data(bd);
> >   	struct drm_device *dev = connector->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	intel_wakeref_t wakeref;
> >   	int ret = 0;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		u32 hw_level;
> >   		drm_modeset_lock(&dev->mode_config.connection_mutex, NULL);
> > @@ -1378,9 +1378,9 @@ void intel_backlight_device_unregister(struct intel_connector *connector)
> >    */
> >   static u32 cnp_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> > -	return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz);
> > +	return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz);
> >   }
> >   /*
> > @@ -1416,7 +1416,7 @@ static u32 spt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >    */
> >   static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 mul, clock;
> > @@ -1425,7 +1425,7 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   	else
> >   		mul = 128;
> > -	if (HAS_PCH_LPT_H(dev_priv))
> > +	if (HAS_PCH_LPT_H(i915))
> >   		clock = MHz(135); /* LPT:H */
> >   	else
> >   		clock = MHz(24); /* LPT:LP */
> > @@ -1439,9 +1439,9 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >    */
> >   static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> > -	return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz * 128);
> > +	return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz * 128);
> >   }
> >   /*
> > @@ -1454,13 +1454,13 @@ static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >    */
> >   static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	int clock;
> > -	if (IS_PINEVIEW(dev_priv))
> > -		clock = KHz(dev_priv->rawclk_freq);
> > +	if (IS_PINEVIEW(i915))
> > +		clock = KHz(i915->rawclk_freq);
> >   	else
> > -		clock = KHz(dev_priv->cdclk.hw.cdclk);
> > +		clock = KHz(i915->cdclk.hw.cdclk);
> >   	return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 32);
> >   }
> > @@ -1472,13 +1472,13 @@ static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >    */
> >   static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	int clock;
> > -	if (IS_G4X(dev_priv))
> > -		clock = KHz(dev_priv->rawclk_freq);
> > +	if (IS_G4X(i915))
> > +		clock = KHz(i915->rawclk_freq);
> >   	else
> > -		clock = KHz(dev_priv->cdclk.hw.cdclk);
> > +		clock = KHz(i915->cdclk.hw.cdclk);
> >   	return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 128);
> >   }
> > @@ -1490,17 +1490,17 @@ static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >    */
> >   static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	int mul, clock;
> >   	if ((I915_READ(CBR1_VLV) & CBR_PWM_CLOCK_MUX_SELECT) == 0) {
> > -		if (IS_CHERRYVIEW(dev_priv))
> > +		if (IS_CHERRYVIEW(i915))
> >   			clock = KHz(19200);
> >   		else
> >   			clock = MHz(25);
> >   		mul = 16;
> >   	} else {
> > -		clock = KHz(dev_priv->rawclk_freq);
> > +		clock = KHz(i915->rawclk_freq);
> >   		mul = 128;
> >   	}
> > @@ -1509,9 +1509,9 @@ static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz)
> >   static u32 get_backlight_max_vbt(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> > -	u16 pwm_freq_hz = dev_priv->vbt.backlight.pwm_freq_hz;
> > +	u16 pwm_freq_hz = i915->vbt.backlight.pwm_freq_hz;
> >   	u32 pwm;
> >   	if (!panel->backlight.hz_to_pwm) {
> > @@ -1542,7 +1542,7 @@ static u32 get_backlight_max_vbt(struct intel_connector *connector)
> >    */
> >   static u32 get_backlight_min_vbt(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	int min;
> > @@ -1555,10 +1555,10 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector)
> >   	 * against this by letting the minimum be at most (arbitrarily chosen)
> >   	 * 25% of the max.
> >   	 */
> > -	min = clamp_t(int, dev_priv->vbt.backlight.min_brightness, 0, 64);
> > -	if (min != dev_priv->vbt.backlight.min_brightness) {
> > +	min = clamp_t(int, i915->vbt.backlight.min_brightness, 0, 64);
> > +	if (min != i915->vbt.backlight.min_brightness) {
> >   		DRM_DEBUG_KMS("clamping VBT min backlight %d/255 to %d/255\n",
> > -			      dev_priv->vbt.backlight.min_brightness, min);
> > +			      i915->vbt.backlight.min_brightness, min);
> >   	}
> >   	/* vbt value is a coefficient in range [0..255] */
> > @@ -1567,12 +1567,12 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector)
> >   static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 cpu_ctl2, pch_ctl1, pch_ctl2, val;
> >   	bool alt, cpu_mode;
> > -	if (HAS_PCH_LPT(dev_priv))
> > +	if (HAS_PCH_LPT(i915))
> >   		alt = I915_READ(SOUTH_CHICKEN2) & LPT_PWM_GRANULARITY;
> >   	else
> >   		alt = I915_READ(SOUTH_CHICKEN1) & SPT_PWM_GRANULARITY;
> > @@ -1596,7 +1596,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus
> >   	panel->backlight.enabled = pch_ctl1 & BLM_PCH_PWM_ENABLE;
> > -	cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(dev_priv) &&
> > +	cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(i915) &&
> >   		   !(pch_ctl1 & BLM_PCH_OVERRIDE_ENABLE) &&
> >   		   (cpu_ctl2 & BLM_PWM_ENABLE);
> >   	if (cpu_mode)
> > @@ -1622,7 +1622,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus
> >   static int pch_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 cpu_ctl2, pch_ctl1, pch_ctl2, val;
> > @@ -1654,16 +1654,16 @@ static int pch_setup_backlight(struct intel_connector *connector, enum pipe unus
> >   static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 ctl, val;
> >   	ctl = I915_READ(BLC_PWM_CTL);
> > -	if (IS_GEN(dev_priv, 2) || IS_I915GM(dev_priv) || IS_I945GM(dev_priv))
> > +	if (IS_GEN(i915, 2) || IS_I915GM(i915) || IS_I945GM(i915))
> >   		panel->backlight.combination_mode = ctl & BLM_LEGACY_MODE;
> > -	if (IS_PINEVIEW(dev_priv))
> > +	if (IS_PINEVIEW(i915))
> >   		panel->backlight.active_low_pwm = ctl & BLM_POLARITY_PNV;
> >   	panel->backlight.max = ctl >> 17;
> > @@ -1693,7 +1693,7 @@ static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unu
> >   static int i965_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 ctl, ctl2, val;
> > @@ -1727,7 +1727,7 @@ static int i965_setup_backlight(struct intel_connector *connector, enum pipe unu
> >   static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 ctl, ctl2, val;
> > @@ -1748,7 +1748,7 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe
> >   	panel->backlight.min = get_backlight_min_vbt(connector);
> > -	val = _vlv_get_backlight(dev_priv, pipe);
> > +	val = _vlv_get_backlight(i915, pipe);
> >   	val = intel_panel_compute_brightness(connector, val);
> >   	panel->backlight.level = clamp(val, panel->backlight.min,
> >   				       panel->backlight.max);
> > @@ -1761,11 +1761,11 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe
> >   static int
> >   bxt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 pwm_ctl, val;
> > -	panel->backlight.controller = dev_priv->vbt.backlight.controller;
> > +	panel->backlight.controller = i915->vbt.backlight.controller;
> >   	pwm_ctl = I915_READ(BXT_BLC_PWM_CTL(panel->backlight.controller));
> > @@ -1801,7 +1801,7 @@ bxt_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   static int
> >   cnp_setup_backlight(struct intel_connector *connector, enum pipe unused)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	u32 pwm_ctl, val;
> > @@ -1881,28 +1881,28 @@ void intel_panel_update_backlight(struct intel_encoder *encoder,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> >   	struct intel_connector *connector = to_intel_connector(conn_state->connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_panel *panel = &connector->panel;
> >   	if (!panel->backlight.present)
> >   		return;
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	if (!panel->backlight.enabled)
> >   		__intel_panel_enable_backlight(crtc_state, conn_state);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   }
> >   int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct intel_connector *intel_connector = to_intel_connector(connector);
> >   	struct intel_panel *panel = &intel_connector->panel;
> >   	int ret;
> > -	if (!dev_priv->vbt.backlight.present) {
> > -		if (dev_priv->quirks & QUIRK_BACKLIGHT_PRESENT) {
> > +	if (!i915->vbt.backlight.present) {
> > +		if (i915->quirks & QUIRK_BACKLIGHT_PRESENT) {
> >   			DRM_DEBUG_KMS("no backlight present per VBT, but present per quirk\n");
> >   		} else {
> >   			DRM_DEBUG_KMS("no backlight present per VBT\n");
> > @@ -1915,9 +1915,9 @@ int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe)
> >   		return -ENODEV;
> >   	/* set level and max in panel struct */
> > -	mutex_lock(&dev_priv->backlight_lock);
> > +	mutex_lock(&i915->backlight_lock);
> >   	ret = panel->backlight.setup(intel_connector, pipe);
> > -	mutex_unlock(&dev_priv->backlight_lock);
> > +	mutex_unlock(&i915->backlight_lock);
> >   	if (ret) {
> >   		DRM_DEBUG_KMS("failed to setup backlight for connector %s\n",
> > @@ -1950,7 +1950,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> >   {
> >   	struct intel_connector *connector =
> >   		container_of(panel, struct intel_connector, panel);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	if (connector->base.connector_type == DRM_MODE_CONNECTOR_eDP &&
> >   	    intel_dp_aux_init_backlight_funcs(connector) == 0)
> > @@ -1960,38 +1960,38 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> >   	    intel_dsi_dcs_init_backlight_funcs(connector) == 0)
> >   		return;
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		panel->backlight.setup = bxt_setup_backlight;
> >   		panel->backlight.enable = bxt_enable_backlight;
> >   		panel->backlight.disable = bxt_disable_backlight;
> >   		panel->backlight.set = bxt_set_backlight;
> >   		panel->backlight.get = bxt_get_backlight;
> >   		panel->backlight.hz_to_pwm = bxt_hz_to_pwm;
> > -	} else if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP) {
> > +	} else if (INTEL_PCH_TYPE(i915) >= PCH_CNP) {
> >   		panel->backlight.setup = cnp_setup_backlight;
> >   		panel->backlight.enable = cnp_enable_backlight;
> >   		panel->backlight.disable = cnp_disable_backlight;
> >   		panel->backlight.set = bxt_set_backlight;
> >   		panel->backlight.get = bxt_get_backlight;
> >   		panel->backlight.hz_to_pwm = cnp_hz_to_pwm;
> > -	} else if (INTEL_PCH_TYPE(dev_priv) >= PCH_LPT) {
> > +	} else if (INTEL_PCH_TYPE(i915) >= PCH_LPT) {
> >   		panel->backlight.setup = lpt_setup_backlight;
> >   		panel->backlight.enable = lpt_enable_backlight;
> >   		panel->backlight.disable = lpt_disable_backlight;
> >   		panel->backlight.set = lpt_set_backlight;
> >   		panel->backlight.get = lpt_get_backlight;
> > -		if (HAS_PCH_LPT(dev_priv))
> > +		if (HAS_PCH_LPT(i915))
> >   			panel->backlight.hz_to_pwm = lpt_hz_to_pwm;
> >   		else
> >   			panel->backlight.hz_to_pwm = spt_hz_to_pwm;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> >   		panel->backlight.setup = pch_setup_backlight;
> >   		panel->backlight.enable = pch_enable_backlight;
> >   		panel->backlight.disable = pch_disable_backlight;
> >   		panel->backlight.set = pch_set_backlight;
> >   		panel->backlight.get = pch_get_backlight;
> >   		panel->backlight.hz_to_pwm = pch_hz_to_pwm;
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		if (connector->base.connector_type == DRM_MODE_CONNECTOR_DSI) {
> >   			panel->backlight.setup = pwm_setup_backlight;
> >   			panel->backlight.enable = pwm_enable_backlight;
> > @@ -2006,7 +2006,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel)
> >   			panel->backlight.get = vlv_get_backlight;
> >   			panel->backlight.hz_to_pwm = vlv_hz_to_pwm;
> >   		}
> > -	} else if (IS_GEN(dev_priv, 4)) {
> > +	} else if (IS_GEN(i915, 4)) {
> >   		panel->backlight.setup = i965_setup_backlight;
> >   		panel->backlight.enable = i965_enable_backlight;
> >   		panel->backlight.disable = i965_disable_backlight;
> > diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.c b/drivers/gpu/drm/i915/intel_pipe_crc.c
> > index 1e2c4307d05a..4a4409f4076f 100644
> > --- a/drivers/gpu/drm/i915/intel_pipe_crc.c
> > +++ b/drivers/gpu/drm/i915/intel_pipe_crc.c
> > @@ -70,11 +70,11 @@ static int i8xx_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source,
> >   	return 0;
> >   }
> > -static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv,
> > +static int i9xx_pipe_crc_auto_source(struct drm_i915_private *i915,
> >   				     enum pipe pipe,
> >   				     enum intel_pipe_crc_source *source)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_encoder *encoder;
> >   	struct intel_crtc *crtc;
> >   	struct intel_digital_port *dig_port;
> > @@ -124,7 +124,7 @@ static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv,
> >   	return ret;
> >   }
> > -static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> > +static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> >   				enum pipe pipe,
> >   				enum intel_pipe_crc_source *source,
> >   				u32 *val)
> > @@ -132,7 +132,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   	bool need_stable_symbols = false;
> >   	if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) {
> > -		int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source);
> > +		int ret = i9xx_pipe_crc_auto_source(i915, pipe, source);
> >   		if (ret)
> >   			return ret;
> >   	}
> > @@ -150,7 +150,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   		need_stable_symbols = true;
> >   		break;
> >   	case INTEL_PIPE_CRC_SOURCE_DP_D:
> > -		if (!IS_CHERRYVIEW(dev_priv))
> > +		if (!IS_CHERRYVIEW(i915))
> >   			return -EINVAL;
> >   		*val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_DP_D_VLV;
> >   		need_stable_symbols = true;
> > @@ -194,13 +194,13 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> > +static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> >   				 enum pipe pipe,
> >   				 enum intel_pipe_crc_source *source,
> >   				 u32 *val)
> >   {
> >   	if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) {
> > -		int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source);
> > +		int ret = i9xx_pipe_crc_auto_source(i915, pipe, source);
> >   		if (ret)
> >   			return ret;
> >   	}
> > @@ -210,7 +210,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   		*val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_PIPE_I9XX;
> >   		break;
> >   	case INTEL_PIPE_CRC_SOURCE_TV:
> > -		if (!SUPPORTS_TV(dev_priv))
> > +		if (!SUPPORTS_TV(i915))
> >   			return -EINVAL;
> >   		*val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_TV_PRE;
> >   		break;
> > @@ -234,7 +234,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *dev_priv,
> > +static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *i915,
> >   					 enum pipe pipe)
> >   {
> >   	u32 tmp = I915_READ(PORT_DFT2_G4X);
> > @@ -286,7 +286,7 @@ static int ilk_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source,
> >   static void
> >   intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_crtc_state *pipe_config;
> >   	struct drm_atomic_state *state;
> >   	struct drm_modeset_acquire_ctx ctx;
> > @@ -294,7 +294,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> >   	drm_modeset_acquire_init(&ctx, 0);
> > -	state = drm_atomic_state_alloc(&dev_priv->drm);
> > +	state = drm_atomic_state_alloc(&i915->drm);
> >   	if (!state) {
> >   		ret = -ENOMEM;
> >   		goto unlock;
> > @@ -312,7 +312,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> >   	pipe_config->base.mode_changed = pipe_config->has_psr;
> >   	pipe_config->crc_enabled = enable;
> > -	if (IS_HASWELL(dev_priv) &&
> > +	if (IS_HASWELL(i915) &&
> >   	    pipe_config->base.active && crtc->pipe == PIPE_A &&
> >   	    pipe_config->cpu_transcoder == TRANSCODER_EDP)
> >   		pipe_config->base.mode_changed = true;
> > @@ -333,7 +333,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable)
> >   	drm_modeset_acquire_fini(&ctx);
> >   }
> > -static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> > +static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> >   				enum pipe pipe,
> >   				enum intel_pipe_crc_source *source,
> >   				u32 *val)
> > @@ -361,7 +361,7 @@ static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> > +static int skl_pipe_crc_ctl_reg(struct drm_i915_private *i915,
> >   				enum pipe pipe,
> >   				enum intel_pipe_crc_source *source,
> >   				u32 *val)
> > @@ -404,22 +404,22 @@ static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv,
> >   	return 0;
> >   }
> > -static int get_new_crc_ctl_reg(struct drm_i915_private *dev_priv,
> > +static int get_new_crc_ctl_reg(struct drm_i915_private *i915,
> >   			       enum pipe pipe,
> >   			       enum intel_pipe_crc_source *source, u32 *val)
> >   {
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		return i8xx_pipe_crc_ctl_reg(source, val);
> > -	else if (INTEL_GEN(dev_priv) < 5)
> > -		return i9xx_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		return vlv_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> > -	else if (IS_GEN_RANGE(dev_priv, 5, 6))
> > +	else if (INTEL_GEN(i915) < 5)
> > +		return i9xx_pipe_crc_ctl_reg(i915, pipe, source, val);
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		return vlv_pipe_crc_ctl_reg(i915, pipe, source, val);
> > +	else if (IS_GEN_RANGE(i915, 5, 6))
> >   		return ilk_pipe_crc_ctl_reg(source, val);
> > -	else if (INTEL_GEN(dev_priv) < 9)
> > -		return ivb_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> > +	else if (INTEL_GEN(i915) < 9)
> > +		return ivb_pipe_crc_ctl_reg(i915, pipe, source, val);
> >   	else
> > -		return skl_pipe_crc_ctl_reg(dev_priv, pipe, source, val);
> > +		return skl_pipe_crc_ctl_reg(i915, pipe, source, val);
> >   }
> >   static int
> > @@ -440,18 +440,18 @@ display_crc_ctl_parse_source(const char *buf, enum intel_pipe_crc_source *s)
> >   	return 0;
> >   }
> > -void intel_display_crc_init(struct drm_i915_private *dev_priv)
> > +void intel_display_crc_init(struct drm_i915_private *i915)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > -		struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe];
> > +	for_each_pipe(i915, pipe) {
> > +		struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe];
> >   		spin_lock_init(&pipe_crc->lock);
> >   	}
> >   }
> > -static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int i8xx_crc_source_valid(struct drm_i915_private *i915,
> >   				 const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -463,7 +463,7 @@ static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int i9xx_crc_source_valid(struct drm_i915_private *i915,
> >   				 const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -476,7 +476,7 @@ static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int vlv_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int vlv_crc_source_valid(struct drm_i915_private *i915,
> >   				const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -491,7 +491,7 @@ static int vlv_crc_source_valid(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int ilk_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int ilk_crc_source_valid(struct drm_i915_private *i915,
> >   				const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -505,7 +505,7 @@ static int ilk_crc_source_valid(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int ivb_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int ivb_crc_source_valid(struct drm_i915_private *i915,
> >   				const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -519,7 +519,7 @@ static int ivb_crc_source_valid(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static int skl_crc_source_valid(struct drm_i915_private *dev_priv,
> > +static int skl_crc_source_valid(struct drm_i915_private *i915,
> >   				const enum intel_pipe_crc_source source)
> >   {
> >   	switch (source) {
> > @@ -539,21 +539,21 @@ static int skl_crc_source_valid(struct drm_i915_private *dev_priv,
> >   }
> >   static int
> > -intel_is_valid_crc_source(struct drm_i915_private *dev_priv,
> > +intel_is_valid_crc_source(struct drm_i915_private *i915,
> >   			  const enum intel_pipe_crc_source source)
> >   {
> > -	if (IS_GEN(dev_priv, 2))
> > -		return i8xx_crc_source_valid(dev_priv, source);
> > -	else if (INTEL_GEN(dev_priv) < 5)
> > -		return i9xx_crc_source_valid(dev_priv, source);
> > -	else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		return vlv_crc_source_valid(dev_priv, source);
> > -	else if (IS_GEN_RANGE(dev_priv, 5, 6))
> > -		return ilk_crc_source_valid(dev_priv, source);
> > -	else if (INTEL_GEN(dev_priv) < 9)
> > -		return ivb_crc_source_valid(dev_priv, source);
> > +	if (IS_GEN(i915, 2))
> > +		return i8xx_crc_source_valid(i915, source);
> > +	else if (INTEL_GEN(i915) < 5)
> > +		return i9xx_crc_source_valid(i915, source);
> > +	else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		return vlv_crc_source_valid(i915, source);
> > +	else if (IS_GEN_RANGE(i915, 5, 6))
> > +		return ilk_crc_source_valid(i915, source);
> > +	else if (INTEL_GEN(i915) < 9)
> > +		return ivb_crc_source_valid(i915, source);
> >   	else
> > -		return skl_crc_source_valid(dev_priv, source);
> > +		return skl_crc_source_valid(i915, source);
> >   }
> >   const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> > @@ -566,7 +566,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> >   int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
> >   				 size_t *values_cnt)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	enum intel_pipe_crc_source source;
> >   	if (display_crc_ctl_parse_source(source_name, &source) < 0) {
> > @@ -575,7 +575,7 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
> >   	}
> >   	if (source == INTEL_PIPE_CRC_SOURCE_AUTO ||
> > -	    intel_is_valid_crc_source(dev_priv, source) == 0) {
> > +	    intel_is_valid_crc_source(i915, source) == 0) {
> >   		*values_cnt = 5;
> >   		return 0;
> >   	}
> > @@ -585,8 +585,8 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name,
> >   int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > -	struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> > +	struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
> >   	enum intel_display_power_domain power_domain;
> >   	enum intel_pipe_crc_source source;
> >   	intel_wakeref_t wakeref;
> > @@ -600,7 +600,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   	}
> >   	power_domain = POWER_DOMAIN_PIPE(crtc->index);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref) {
> >   		DRM_DEBUG_KMS("Trying to capture CRC while pipe is off\n");
> >   		return -EIO;
> > @@ -610,7 +610,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   	if (enable)
> >   		intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), true);
> > -	ret = get_new_crc_ctl_reg(dev_priv, crtc->index, &source, &val);
> > +	ret = get_new_crc_ctl_reg(i915, crtc->index, &source, &val);
> >   	if (ret != 0)
> >   		goto out;
> > @@ -619,8 +619,8 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   	POSTING_READ(PIPE_CRC_CTL(crtc->index));
> >   	if (!source) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -			vlv_undo_pipe_scramble_reset(dev_priv, crtc->index);
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +			vlv_undo_pipe_scramble_reset(i915, crtc->index);
> >   	}
> >   	pipe_crc->skipped = 0;
> > @@ -629,7 +629,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   	if (!enable)
> >   		intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), false);
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -637,14 +637,14 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name)
> >   void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc)
> >   {
> >   	struct drm_crtc *crtc = &intel_crtc->base;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > -	struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> > +	struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
> >   	u32 val = 0;
> >   	if (!crtc->crc.opened)
> >   		return;
> > -	if (get_new_crc_ctl_reg(dev_priv, crtc->index, &pipe_crc->source, &val) < 0)
> > +	if (get_new_crc_ctl_reg(i915, crtc->index, &pipe_crc->source, &val) < 0)
> >   		return;
> >   	/* Don't need pipe_crc->lock here, IRQs are not generated. */
> > @@ -657,8 +657,8 @@ void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc)
> >   void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc)
> >   {
> >   	struct drm_crtc *crtc = &intel_crtc->base;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > -	struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index];
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> > +	struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index];
> >   	/* Swallow crc's until we stop generating them. */
> >   	spin_lock_irq(&pipe_crc->lock);
> > @@ -667,5 +667,5 @@ void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc)
> >   	I915_WRITE(PIPE_CRC_CTL(crtc->index), 0);
> >   	POSTING_READ(PIPE_CRC_CTL(crtc->index));
> > -	synchronize_irq(dev_priv->drm.irq);
> > +	synchronize_irq(i915->drm.irq);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.h b/drivers/gpu/drm/i915/intel_pipe_crc.h
> > index db258a756fc6..42048ab1391a 100644
> > --- a/drivers/gpu/drm/i915/intel_pipe_crc.h
> > +++ b/drivers/gpu/drm/i915/intel_pipe_crc.h
> > @@ -13,7 +13,7 @@ struct drm_i915_private;
> >   struct intel_crtc;
> >   #ifdef CONFIG_DEBUG_FS
> > -void intel_display_crc_init(struct drm_i915_private *dev_priv);
> > +void intel_display_crc_init(struct drm_i915_private *i915);
> >   int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name);
> >   int intel_crtc_verify_crc_source(struct drm_crtc *crtc,
> >   				 const char *source_name, size_t *values_cnt);
> > @@ -22,7 +22,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc,
> >   void intel_crtc_disable_pipe_crc(struct intel_crtc *crtc);
> >   void intel_crtc_enable_pipe_crc(struct intel_crtc *crtc);
> >   #else
> > -static inline void intel_display_crc_init(struct drm_i915_private *dev_priv) {}
> > +static inline void intel_display_crc_init(struct drm_i915_private *i915) {}
> >   #define intel_crtc_set_crc_source NULL
> >   #define intel_crtc_verify_crc_source NULL
> >   #define intel_crtc_get_crc_sources NULL
> > diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c
> > index 2c7f3ebc0117..4b939160cd95 100644
> > --- a/drivers/gpu/drm/i915/intel_pm.c
> > +++ b/drivers/gpu/drm/i915/intel_pm.c
> > @@ -63,9 +63,9 @@
> >    * require higher latency to switch to and wake up.
> >    */
> > -static void gen9_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void gen9_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	if (HAS_LLC(dev_priv)) {
> > +	if (HAS_LLC(i915)) {
> >   		/*
> >   		 * WaCompressedResourceDisplayNewHashMode:skl,kbl
> >   		 * Display WA #0390: skl,kbl
> > @@ -96,16 +96,16 @@ static void gen9_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) |
> >   		   ILK_DPFC_DISABLE_DUMMY0);
> > -	if (IS_SKYLAKE(dev_priv)) {
> > +	if (IS_SKYLAKE(i915)) {
> >   		/* WaDisableDopClockGating */
> >   		I915_WRITE(GEN7_MISCCPCTL, I915_READ(GEN7_MISCCPCTL)
> >   			   & ~GEN7_DOP_CLOCK_GATE_ENABLE);
> >   	}
> >   }
> > -static void bxt_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void bxt_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	gen9_init_clock_gating(dev_priv);
> > +	gen9_init_clock_gating(i915);
> >   	/* WaDisableSDEUnitClockGating:bxt */
> >   	I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) |
> > @@ -126,9 +126,9 @@ static void bxt_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   PWM1_GATING_DIS | PWM2_GATING_DIS);
> >   }
> > -static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void glk_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	gen9_init_clock_gating(dev_priv);
> > +	gen9_init_clock_gating(i915);
> >   	/*
> >   	 * WaDisablePWMClockGating:glk
> > @@ -139,7 +139,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   PWM1_GATING_DIS | PWM2_GATING_DIS);
> >   	/* WaDDIIOTimeout:glk */
> > -	if (IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1)) {
> > +	if (IS_GLK_REVID(i915, 0, GLK_REVID_A1)) {
> >   		u32 val = I915_READ(CHICKEN_MISC_2);
> >   		val &= ~(GLK_CL0_PWR_DOWN |
> >   			 GLK_CL1_PWR_DOWN |
> > @@ -149,7 +149,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv)
> >   }
> > -static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv)
> > +static void i915_pineview_get_mem_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 tmp;
> > @@ -157,100 +157,100 @@ static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv)
> >   	switch (tmp & CLKCFG_FSB_MASK) {
> >   	case CLKCFG_FSB_533:
> > -		dev_priv->fsb_freq = 533; /* 133*4 */
> > +		i915->fsb_freq = 533; /* 133*4 */
> >   		break;
> >   	case CLKCFG_FSB_800:
> > -		dev_priv->fsb_freq = 800; /* 200*4 */
> > +		i915->fsb_freq = 800; /* 200*4 */
> >   		break;
> >   	case CLKCFG_FSB_667:
> > -		dev_priv->fsb_freq =  667; /* 167*4 */
> > +		i915->fsb_freq =  667; /* 167*4 */
> >   		break;
> >   	case CLKCFG_FSB_400:
> > -		dev_priv->fsb_freq = 400; /* 100*4 */
> > +		i915->fsb_freq = 400; /* 100*4 */
> >   		break;
> >   	}
> >   	switch (tmp & CLKCFG_MEM_MASK) {
> >   	case CLKCFG_MEM_533:
> > -		dev_priv->mem_freq = 533;
> > +		i915->mem_freq = 533;
> >   		break;
> >   	case CLKCFG_MEM_667:
> > -		dev_priv->mem_freq = 667;
> > +		i915->mem_freq = 667;
> >   		break;
> >   	case CLKCFG_MEM_800:
> > -		dev_priv->mem_freq = 800;
> > +		i915->mem_freq = 800;
> >   		break;
> >   	}
> >   	/* detect pineview DDR3 setting */
> >   	tmp = I915_READ(CSHRDDR3CTL);
> > -	dev_priv->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0;
> > +	i915->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0;
> >   }
> > -static void i915_ironlake_get_mem_freq(struct drm_i915_private *dev_priv)
> > +static void i915_ironlake_get_mem_freq(struct drm_i915_private *i915)
> >   {
> >   	u16 ddrpll, csipll;
> > -	ddrpll = intel_uncore_read16(&dev_priv->uncore, DDRMPLL1);
> > -	csipll = intel_uncore_read16(&dev_priv->uncore, CSIPLL0);
> > +	ddrpll = intel_uncore_read16(&i915->uncore, DDRMPLL1);
> > +	csipll = intel_uncore_read16(&i915->uncore, CSIPLL0);
> >   	switch (ddrpll & 0xff) {
> >   	case 0xc:
> > -		dev_priv->mem_freq = 800;
> > +		i915->mem_freq = 800;
> >   		break;
> >   	case 0x10:
> > -		dev_priv->mem_freq = 1066;
> > +		i915->mem_freq = 1066;
> >   		break;
> >   	case 0x14:
> > -		dev_priv->mem_freq = 1333;
> > +		i915->mem_freq = 1333;
> >   		break;
> >   	case 0x18:
> > -		dev_priv->mem_freq = 1600;
> > +		i915->mem_freq = 1600;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_DRIVER("unknown memory frequency 0x%02x\n",
> >   				 ddrpll & 0xff);
> > -		dev_priv->mem_freq = 0;
> > +		i915->mem_freq = 0;
> >   		break;
> >   	}
> > -	dev_priv->ips.r_t = dev_priv->mem_freq;
> > +	i915->ips.r_t = i915->mem_freq;
> >   	switch (csipll & 0x3ff) {
> >   	case 0x00c:
> > -		dev_priv->fsb_freq = 3200;
> > +		i915->fsb_freq = 3200;
> >   		break;
> >   	case 0x00e:
> > -		dev_priv->fsb_freq = 3733;
> > +		i915->fsb_freq = 3733;
> >   		break;
> >   	case 0x010:
> > -		dev_priv->fsb_freq = 4266;
> > +		i915->fsb_freq = 4266;
> >   		break;
> >   	case 0x012:
> > -		dev_priv->fsb_freq = 4800;
> > +		i915->fsb_freq = 4800;
> >   		break;
> >   	case 0x014:
> > -		dev_priv->fsb_freq = 5333;
> > +		i915->fsb_freq = 5333;
> >   		break;
> >   	case 0x016:
> > -		dev_priv->fsb_freq = 5866;
> > +		i915->fsb_freq = 5866;
> >   		break;
> >   	case 0x018:
> > -		dev_priv->fsb_freq = 6400;
> > +		i915->fsb_freq = 6400;
> >   		break;
> >   	default:
> >   		DRM_DEBUG_DRIVER("unknown fsb frequency 0x%04x\n",
> >   				 csipll & 0x3ff);
> > -		dev_priv->fsb_freq = 0;
> > +		i915->fsb_freq = 0;
> >   		break;
> >   	}
> > -	if (dev_priv->fsb_freq == 3200) {
> > -		dev_priv->ips.c_m = 0;
> > -	} else if (dev_priv->fsb_freq > 3200 && dev_priv->fsb_freq <= 4800) {
> > -		dev_priv->ips.c_m = 1;
> > +	if (i915->fsb_freq == 3200) {
> > +		i915->ips.c_m = 0;
> > +	} else if (i915->fsb_freq > 3200 && i915->fsb_freq <= 4800) {
> > +		i915->ips.c_m = 1;
> >   	} else {
> > -		dev_priv->ips.c_m = 2;
> > +		i915->ips.c_m = 2;
> >   	}
> >   }
> > @@ -316,61 +316,61 @@ static const struct cxsr_latency *intel_get_cxsr_latency(bool is_desktop,
> >   	return NULL;
> >   }
> > -static void chv_set_memory_dvfs(struct drm_i915_private *dev_priv, bool enable)
> > +static void chv_set_memory_dvfs(struct drm_i915_private *i915, bool enable)
> >   {
> >   	u32 val;
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> > +	val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> >   	if (enable)
> >   		val &= ~FORCE_DDR_HIGH_FREQ;
> >   	else
> >   		val |= FORCE_DDR_HIGH_FREQ;
> >   	val &= ~FORCE_DDR_LOW_FREQ;
> >   	val |= FORCE_DDR_FREQ_REQ_ACK;
> > -	vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val);
> > +	vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val);
> > -	if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) &
> > +	if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) &
> >   		      FORCE_DDR_FREQ_REQ_ACK) == 0, 3))
> >   		DRM_ERROR("timed out waiting for Punit DDR DVFS request\n");
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   }
> > -static void chv_set_memory_pm5(struct drm_i915_private *dev_priv, bool enable)
> > +static void chv_set_memory_pm5(struct drm_i915_private *i915, bool enable)
> >   {
> >   	u32 val;
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +	val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> >   	if (enable)
> >   		val |= DSP_MAXFIFO_PM5_ENABLE;
> >   	else
> >   		val &= ~DSP_MAXFIFO_PM5_ENABLE;
> > -	vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val);
> > +	vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val);
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   }
> >   #define FW_WM(value, plane) \
> >   	(((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK)
> > -static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable)
> > +static bool _intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable)
> >   {
> >   	bool was_enabled;
> >   	u32 val;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		was_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
> >   		I915_WRITE(FW_BLC_SELF_VLV, enable ? FW_CSPWRDWNEN : 0);
> >   		POSTING_READ(FW_BLC_SELF_VLV);
> > -	} else if (IS_G4X(dev_priv) || IS_I965GM(dev_priv)) {
> > +	} else if (IS_G4X(i915) || IS_I965GM(i915)) {
> >   		was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> >   		I915_WRITE(FW_BLC_SELF, enable ? FW_BLC_SELF_EN : 0);
> >   		POSTING_READ(FW_BLC_SELF);
> > -	} else if (IS_PINEVIEW(dev_priv)) {
> > +	} else if (IS_PINEVIEW(i915)) {
> >   		val = I915_READ(DSPFW3);
> >   		was_enabled = val & PINEVIEW_SELF_REFRESH_EN;
> >   		if (enable)
> > @@ -379,13 +379,13 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> >   			val &= ~PINEVIEW_SELF_REFRESH_EN;
> >   		I915_WRITE(DSPFW3, val);
> >   		POSTING_READ(DSPFW3);
> > -	} else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv)) {
> > +	} else if (IS_I945G(i915) || IS_I945GM(i915)) {
> >   		was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> >   		val = enable ? _MASKED_BIT_ENABLE(FW_BLC_SELF_EN) :
> >   			       _MASKED_BIT_DISABLE(FW_BLC_SELF_EN);
> >   		I915_WRITE(FW_BLC_SELF, val);
> >   		POSTING_READ(FW_BLC_SELF);
> > -	} else if (IS_I915GM(dev_priv)) {
> > +	} else if (IS_I915GM(i915)) {
> >   		/*
> >   		 * FIXME can't find a bit like this for 915G, and
> >   		 * and yet it does have the related watermark in
> > @@ -400,7 +400,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> >   		return false;
> >   	}
> > -	trace_intel_memory_cxsr(dev_priv, was_enabled, enable);
> > +	trace_intel_memory_cxsr(i915, was_enabled, enable);
> >   	DRM_DEBUG_KMS("memory self-refresh is %s (was %s)\n",
> >   		      enableddisabled(enable),
> > @@ -411,7 +411,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> >   /**
> >    * intel_set_memory_cxsr - Configure CxSR state
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @enable: Allow vs. disallow CxSR
> >    *
> >    * Allow or disallow the system to enter a special CxSR
> > @@ -446,17 +446,17 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl
> >    * the hardware w.r.t. HPLL SR when writing to plane registers.
> >    * Disallowing just CxSR is sufficient.
> >    */
> > -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable)
> > +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable)
> >   {
> >   	bool ret;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > -	ret = _intel_set_memory_cxsr(dev_priv, enable);
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->wm.vlv.cxsr = enable;
> > -	else if (IS_G4X(dev_priv))
> > -		dev_priv->wm.g4x.cxsr = enable;
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> > +	ret = _intel_set_memory_cxsr(i915, enable);
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		i915->wm.vlv.cxsr = enable;
> > +	else if (IS_G4X(i915))
> > +		i915->wm.g4x.cxsr = enable;
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   	return ret;
> >   }
> > @@ -483,7 +483,7 @@ static const int pessimal_latency_ns = 5000;
> >   static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct vlv_fifo_state *fifo_state = &crtc_state->wm.vlv.fifo_state;
> >   	enum pipe pipe = crtc->pipe;
> >   	int sprite0_start, sprite1_start;
> > @@ -519,7 +519,7 @@ static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state)
> >   	fifo_state->plane[PLANE_CURSOR] = 63;
> >   }
> > -static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv,
> > +static int i9xx_get_fifo_size(struct drm_i915_private *i915,
> >   			      enum i9xx_plane_id i9xx_plane)
> >   {
> >   	u32 dsparb = I915_READ(DSPARB);
> > @@ -535,7 +535,7 @@ static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv,
> >   	return size;
> >   }
> > -static int i830_get_fifo_size(struct drm_i915_private *dev_priv,
> > +static int i830_get_fifo_size(struct drm_i915_private *i915,
> >   			      enum i9xx_plane_id i9xx_plane)
> >   {
> >   	u32 dsparb = I915_READ(DSPARB);
> > @@ -552,7 +552,7 @@ static int i830_get_fifo_size(struct drm_i915_private *dev_priv,
> >   	return size;
> >   }
> > -static int i845_get_fifo_size(struct drm_i915_private *dev_priv,
> > +static int i845_get_fifo_size(struct drm_i915_private *i915,
> >   			      enum i9xx_plane_id i9xx_plane)
> >   {
> >   	u32 dsparb = I915_READ(DSPARB);
> > @@ -805,9 +805,9 @@ static bool is_enabling(int old, int new, int threshold)
> >   	return old < threshold && new >= threshold;
> >   }
> > -static int intel_wm_num_levels(struct drm_i915_private *dev_priv)
> > +static int intel_wm_num_levels(struct drm_i915_private *i915)
> >   {
> > -	return dev_priv->wm.max_level + 1;
> > +	return i915->wm.max_level + 1;
> >   }
> >   static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state,
> > @@ -833,11 +833,11 @@ static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state,
> >   		return plane_state->base.visible;
> >   }
> > -static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv)
> > +static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *i915)
> >   {
> >   	struct intel_crtc *crtc, *enabled = NULL;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		if (intel_crtc_active(crtc)) {
> >   			if (enabled)
> >   				return NULL;
> > @@ -850,23 +850,23 @@ static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv)
> >   static void pineview_update_wm(struct intel_crtc *unused_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> >   	struct intel_crtc *crtc;
> >   	const struct cxsr_latency *latency;
> >   	u32 reg;
> >   	unsigned int wm;
> > -	latency = intel_get_cxsr_latency(!IS_MOBILE(dev_priv),
> > -					 dev_priv->is_ddr3,
> > -					 dev_priv->fsb_freq,
> > -					 dev_priv->mem_freq);
> > +	latency = intel_get_cxsr_latency(!IS_MOBILE(i915),
> > +					 i915->is_ddr3,
> > +					 i915->fsb_freq,
> > +					 i915->mem_freq);
> >   	if (!latency) {
> >   		DRM_DEBUG_KMS("Unknown FSB/MEM found, disable CxSR\n");
> > -		intel_set_memory_cxsr(dev_priv, false);
> > +		intel_set_memory_cxsr(i915, false);
> >   		return;
> >   	}
> > -	crtc = single_enabled_crtc(dev_priv);
> > +	crtc = single_enabled_crtc(i915);
> >   	if (crtc) {
> >   		const struct drm_display_mode *adjusted_mode =
> >   			&crtc->config->base.adjusted_mode;
> > @@ -913,9 +913,9 @@ static void pineview_update_wm(struct intel_crtc *unused_crtc)
> >   		I915_WRITE(DSPFW3, reg);
> >   		DRM_DEBUG_KMS("DSPFW3 register is %x\n", reg);
> > -		intel_set_memory_cxsr(dev_priv, true);
> > +		intel_set_memory_cxsr(i915, true);
> >   	} else {
> > -		intel_set_memory_cxsr(dev_priv, false);
> > +		intel_set_memory_cxsr(i915, false);
> >   	}
> >   }
> > @@ -936,13 +936,13 @@ static unsigned int g4x_tlb_miss_wa(int fifo_size, int width, int cpp)
> >   	return max(0, tlb_miss);
> >   }
> > -static void g4x_write_wm_values(struct drm_i915_private *dev_priv,
> > +static void g4x_write_wm_values(struct drm_i915_private *i915,
> >   				const struct g4x_wm_values *wm)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe)
> > -		trace_g4x_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm);
> > +	for_each_pipe(i915, pipe)
> > +		trace_g4x_wm(intel_get_crtc_for_pipe(i915, pipe), wm);
> >   	I915_WRITE(DSPFW1,
> >   		   FW_WM(wm->sr.plane, SR) |
> > @@ -968,13 +968,13 @@ static void g4x_write_wm_values(struct drm_i915_private *dev_priv,
> >   #define FW_WM_VLV(value, plane) \
> >   	(((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK_VLV)
> > -static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
> > +static void vlv_write_wm_values(struct drm_i915_private *i915,
> >   				const struct vlv_wm_values *wm)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > -		trace_vlv_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm);
> > +	for_each_pipe(i915, pipe) {
> > +		trace_vlv_wm(intel_get_crtc_for_pipe(i915, pipe), wm);
> >   		I915_WRITE(VLV_DDL(pipe),
> >   			   (wm->ddl[pipe].plane[PLANE_CURSOR] << DDL_CURSOR_SHIFT) |
> > @@ -1006,7 +1006,7 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(DSPFW3,
> >   		   FW_WM(wm->sr.cursor, CURSOR_SR));
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		I915_WRITE(DSPFW7_CHV,
> >   			   FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE1], SPRITED) |
> >   			   FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE0], SPRITEC));
> > @@ -1046,14 +1046,14 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv,
> >   #undef FW_WM_VLV
> > -static void g4x_setup_wm_latency(struct drm_i915_private *dev_priv)
> > +static void g4x_setup_wm_latency(struct drm_i915_private *i915)
> >   {
> >   	/* all latencies in usec */
> > -	dev_priv->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
> > -	dev_priv->wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
> > -	dev_priv->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
> > +	i915->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
> > +	i915->wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
> > +	i915->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
> > -	dev_priv->wm.max_level = G4X_WM_LEVEL_HPLL;
> > +	i915->wm.max_level = G4X_WM_LEVEL_HPLL;
> >   }
> >   static int g4x_plane_fifo_size(enum plane_id plane_id, int level)
> > @@ -1103,10 +1103,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> >   			  int level)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_display_mode *adjusted_mode =
> >   		&crtc_state->base.adjusted_mode;
> > -	unsigned int latency = dev_priv->wm.pri_latency[level] * 10;
> > +	unsigned int latency = i915->wm.pri_latency[level] * 10;
> >   	unsigned int clock, htotal, cpp, width, wm;
> >   	if (latency == 0)
> > @@ -1126,7 +1126,7 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> >   	 * the HPLL watermark, which seems a little strange.
> >   	 * Let's use 32bpp for the HPLL watermark as well.
> >   	 */
> > -	if (IS_GM45(dev_priv) && plane->id == PLANE_PRIMARY &&
> > +	if (IS_GM45(i915) && plane->id == PLANE_PRIMARY &&
> >   	    level != G4X_WM_LEVEL_NORMAL)
> >   		cpp = 4;
> >   	else
> > @@ -1165,10 +1165,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state,
> >   static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> >   				 int level, enum plane_id plane_id, u16 value)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	bool dirty = false;
> > -	for (; level < intel_wm_num_levels(dev_priv); level++) {
> > +	for (; level < intel_wm_num_levels(i915); level++) {
> >   		struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level];
> >   		dirty |= raw->plane[plane_id] != value;
> > @@ -1181,13 +1181,13 @@ static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> >   static bool g4x_raw_fbc_wm_set(struct intel_crtc_state *crtc_state,
> >   			       int level, u16 value)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	bool dirty = false;
> >   	/* NORMAL level doesn't have an FBC watermark */
> >   	level = max(level, G4X_WM_LEVEL_SR);
> > -	for (; level < intel_wm_num_levels(dev_priv); level++) {
> > +	for (; level < intel_wm_num_levels(i915); level++) {
> >   		struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level];
> >   		dirty |= raw->fbc != value;
> > @@ -1283,9 +1283,9 @@ static bool g4x_raw_plane_wm_is_valid(const struct intel_crtc_state *crtc_state,
> >   static bool g4x_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state,
> >   				     int level)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > -	if (level > dev_priv->wm.max_level)
> > +	if (level > i915->wm.max_level)
> >   		return false;
> >   	return g4x_raw_plane_wm_is_valid(crtc_state, PLANE_PRIMARY, level) &&
> > @@ -1484,7 +1484,7 @@ static int g4x_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state)
> >   	return 0;
> >   }
> > -static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> > +static void g4x_merge_wm(struct drm_i915_private *i915,
> >   			 struct g4x_wm_values *wm)
> >   {
> >   	struct intel_crtc *crtc;
> > @@ -1494,7 +1494,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> >   	wm->hpll_en = true;
> >   	wm->fbc_en = true;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x;
> >   		if (!crtc->active)
> > @@ -1516,7 +1516,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> >   		wm->fbc_en = false;
> >   	}
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x;
> >   		enum pipe pipe = crtc->pipe;
> > @@ -1528,23 +1528,23 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void g4x_program_watermarks(struct drm_i915_private *dev_priv)
> > +static void g4x_program_watermarks(struct drm_i915_private *i915)
> >   {
> > -	struct g4x_wm_values *old_wm = &dev_priv->wm.g4x;
> > +	struct g4x_wm_values *old_wm = &i915->wm.g4x;
> >   	struct g4x_wm_values new_wm = {};
> > -	g4x_merge_wm(dev_priv, &new_wm);
> > +	g4x_merge_wm(i915, &new_wm);
> >   	if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0)
> >   		return;
> >   	if (is_disabling(old_wm->cxsr, new_wm.cxsr, true))
> > -		_intel_set_memory_cxsr(dev_priv, false);
> > +		_intel_set_memory_cxsr(i915, false);
> > -	g4x_write_wm_values(dev_priv, &new_wm);
> > +	g4x_write_wm_values(i915, &new_wm);
> >   	if (is_enabling(old_wm->cxsr, new_wm.cxsr, true))
> > -		_intel_set_memory_cxsr(dev_priv, true);
> > +		_intel_set_memory_cxsr(i915, true);
> >   	*old_wm = new_wm;
> >   }
> > @@ -1552,28 +1552,28 @@ static void g4x_program_watermarks(struct drm_i915_private *dev_priv)
> >   static void g4x_initial_watermarks(struct intel_atomic_state *state,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	crtc->wm.active.g4x = crtc_state->wm.g4x.intermediate;
> > -	g4x_program_watermarks(dev_priv);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	g4x_program_watermarks(i915);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   static void g4x_optimize_watermarks(struct intel_atomic_state *state,
> >   				    struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	if (!crtc_state->wm.need_postvbl_update)
> >   		return;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	intel_crtc->wm.active.g4x = crtc_state->wm.g4x.optimal;
> > -	g4x_program_watermarks(dev_priv);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	g4x_program_watermarks(i915);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   /* latency must be in 0.1us units. */
> > @@ -1592,18 +1592,18 @@ static unsigned int vlv_wm_method2(unsigned int pixel_rate,
> >   	return ret;
> >   }
> > -static void vlv_setup_wm_latency(struct drm_i915_private *dev_priv)
> > +static void vlv_setup_wm_latency(struct drm_i915_private *i915)
> >   {
> >   	/* all latencies in usec */
> > -	dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
> > +	i915->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
> > -	dev_priv->wm.max_level = VLV_WM_LEVEL_PM2;
> > +	i915->wm.max_level = VLV_WM_LEVEL_PM2;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > -		dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
> > -		dev_priv->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
> > +	if (IS_CHERRYVIEW(i915)) {
> > +		i915->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
> > +		i915->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
> > -		dev_priv->wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
> > +		i915->wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
> >   	}
> >   }
> > @@ -1612,12 +1612,12 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state,
> >   				int level)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_display_mode *adjusted_mode =
> >   		&crtc_state->base.adjusted_mode;
> >   	unsigned int clock, htotal, cpp, width, wm;
> > -	if (dev_priv->wm.pri_latency[level] == 0)
> > +	if (i915->wm.pri_latency[level] == 0)
> >   		return USHRT_MAX;
> >   	if (!intel_wm_plane_visible(crtc_state, plane_state))
> > @@ -1638,7 +1638,7 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state,
> >   		wm = 63;
> >   	} else {
> >   		wm = vlv_wm_method2(clock, htotal, width, cpp,
> > -				    dev_priv->wm.pri_latency[level] * 10);
> > +				    i915->wm.pri_latency[level] * 10);
> >   	}
> >   	return min_t(unsigned int, wm, USHRT_MAX);
> > @@ -1736,9 +1736,9 @@ static int vlv_compute_fifo(struct intel_crtc_state *crtc_state)
> >   static void vlv_invalidate_wms(struct intel_crtc *crtc,
> >   			       struct vlv_wm_state *wm_state, int level)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	for (; level < intel_wm_num_levels(dev_priv); level++) {
> > +	for (; level < intel_wm_num_levels(i915); level++) {
> >   		enum plane_id plane_id;
> >   		for_each_plane_id_on_crtc(crtc, plane_id)
> > @@ -1764,8 +1764,8 @@ static u16 vlv_invert_wm_value(u16 wm, u16 fifo_size)
> >   static bool vlv_raw_plane_wm_set(struct intel_crtc_state *crtc_state,
> >   				 int level, enum plane_id plane_id, u16 value)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > -	int num_levels = intel_wm_num_levels(dev_priv);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > +	int num_levels = intel_wm_num_levels(i915);
> >   	bool dirty = false;
> >   	for (; level < num_levels; level++) {
> > @@ -1840,7 +1840,7 @@ static bool vlv_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state,
> >   static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_atomic_state *state =
> >   		to_intel_atomic_state(crtc_state->base.state);
> >   	struct vlv_wm_state *wm_state = &crtc_state->wm.vlv.optimal;
> > @@ -1897,7 +1897,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
> >   	}
> >   	/* initially allow all levels */
> > -	wm_state->num_levels = intel_wm_num_levels(dev_priv);
> > +	wm_state->num_levels = intel_wm_num_levels(i915);
> >   	/*
> >   	 * Note that enabling cxsr with no primary/sprite planes
> >   	 * enabled can wedge the pipe. Hence we only allow cxsr
> > @@ -1907,7 +1907,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state)
> >   	for (level = 0; level < wm_state->num_levels; level++) {
> >   		const struct g4x_pipe_wm *raw = &crtc_state->wm.vlv.raw[level];
> > -		const int sr_fifo_size = INTEL_INFO(dev_priv)->num_pipes * 512 - 1;
> > +		const int sr_fifo_size = INTEL_INFO(i915)->num_pipes * 512 - 1;
> >   		if (!vlv_raw_crtc_wm_is_valid(crtc_state, level))
> >   			break;
> > @@ -1948,8 +1948,8 @@ static void vlv_atomic_update_fifo(struct intel_atomic_state *state,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	const struct vlv_fifo_state *fifo_state =
> >   		&crtc_state->wm.vlv.fifo_state;
> >   	int sprite0_start, sprite1_start, fifo_size;
> > @@ -2092,16 +2092,16 @@ static int vlv_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state)
> >   	return 0;
> >   }
> > -static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> > +static void vlv_merge_wm(struct drm_i915_private *i915,
> >   			 struct vlv_wm_values *wm)
> >   {
> >   	struct intel_crtc *crtc;
> >   	int num_active_crtcs = 0;
> > -	wm->level = dev_priv->wm.max_level;
> > +	wm->level = i915->wm.max_level;
> >   	wm->cxsr = true;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv;
> >   		if (!crtc->active)
> > @@ -2120,7 +2120,7 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> >   	if (num_active_crtcs > 1)
> >   		wm->level = VLV_WM_LEVEL_PM2;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv;
> >   		enum pipe pipe = crtc->pipe;
> > @@ -2135,35 +2135,35 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void vlv_program_watermarks(struct drm_i915_private *dev_priv)
> > +static void vlv_program_watermarks(struct drm_i915_private *i915)
> >   {
> > -	struct vlv_wm_values *old_wm = &dev_priv->wm.vlv;
> > +	struct vlv_wm_values *old_wm = &i915->wm.vlv;
> >   	struct vlv_wm_values new_wm = {};
> > -	vlv_merge_wm(dev_priv, &new_wm);
> > +	vlv_merge_wm(i915, &new_wm);
> >   	if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0)
> >   		return;
> >   	if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS))
> > -		chv_set_memory_dvfs(dev_priv, false);
> > +		chv_set_memory_dvfs(i915, false);
> >   	if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5))
> > -		chv_set_memory_pm5(dev_priv, false);
> > +		chv_set_memory_pm5(i915, false);
> >   	if (is_disabling(old_wm->cxsr, new_wm.cxsr, true))
> > -		_intel_set_memory_cxsr(dev_priv, false);
> > +		_intel_set_memory_cxsr(i915, false);
> > -	vlv_write_wm_values(dev_priv, &new_wm);
> > +	vlv_write_wm_values(i915, &new_wm);
> >   	if (is_enabling(old_wm->cxsr, new_wm.cxsr, true))
> > -		_intel_set_memory_cxsr(dev_priv, true);
> > +		_intel_set_memory_cxsr(i915, true);
> >   	if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5))
> > -		chv_set_memory_pm5(dev_priv, true);
> > +		chv_set_memory_pm5(i915, true);
> >   	if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS))
> > -		chv_set_memory_dvfs(dev_priv, true);
> > +		chv_set_memory_dvfs(i915, true);
> >   	*old_wm = new_wm;
> >   }
> > @@ -2171,40 +2171,40 @@ static void vlv_program_watermarks(struct drm_i915_private *dev_priv)
> >   static void vlv_initial_watermarks(struct intel_atomic_state *state,
> >   				   struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	crtc->wm.active.vlv = crtc_state->wm.vlv.intermediate;
> > -	vlv_program_watermarks(dev_priv);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	vlv_program_watermarks(i915);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   static void vlv_optimize_watermarks(struct intel_atomic_state *state,
> >   				    struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	if (!crtc_state->wm.need_postvbl_update)
> >   		return;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	intel_crtc->wm.active.vlv = crtc_state->wm.vlv.optimal;
> > -	vlv_program_watermarks(dev_priv);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	vlv_program_watermarks(i915);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   static void i965_update_wm(struct intel_crtc *unused_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> >   	struct intel_crtc *crtc;
> >   	int srwm = 1;
> >   	int cursor_sr = 16;
> >   	bool cxsr_enabled;
> >   	/* Calc sr entries for one plane configs */
> > -	crtc = single_enabled_crtc(dev_priv);
> > +	crtc = single_enabled_crtc(i915);
> >   	if (crtc) {
> >   		/* self-refresh has much higher latency */
> >   		static const int sr_latency_ns = 12000;
> > @@ -2246,7 +2246,7 @@ static void i965_update_wm(struct intel_crtc *unused_crtc)
> >   	} else {
> >   		cxsr_enabled = false;
> >   		/* Turn off self refresh if both pipes are enabled */
> > -		intel_set_memory_cxsr(dev_priv, false);
> > +		intel_set_memory_cxsr(i915, false);
> >   	}
> >   	DRM_DEBUG_KMS("Setting FIFO watermarks - A: 8, B: 8, C: 8, SR %d\n",
> > @@ -2263,14 +2263,14 @@ static void i965_update_wm(struct intel_crtc *unused_crtc)
> >   	I915_WRITE(DSPFW3, FW_WM(cursor_sr, CURSOR_SR));
> >   	if (cxsr_enabled)
> > -		intel_set_memory_cxsr(dev_priv, true);
> > +		intel_set_memory_cxsr(i915, true);
> >   }
> >   #undef FW_WM
> >   static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> >   	const struct intel_watermark_params *wm_info;
> >   	u32 fwater_lo;
> >   	u32 fwater_hi;
> > @@ -2279,15 +2279,15 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   	int planea_wm, planeb_wm;
> >   	struct intel_crtc *crtc, *enabled = NULL;
> > -	if (IS_I945GM(dev_priv))
> > +	if (IS_I945GM(i915))
> >   		wm_info = &i945_wm_info;
> > -	else if (!IS_GEN(dev_priv, 2))
> > +	else if (!IS_GEN(i915, 2))
> >   		wm_info = &i915_wm_info;
> >   	else
> >   		wm_info = &i830_a_wm_info;
> > -	fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_A);
> > -	crtc = intel_get_crtc_for_plane(dev_priv, PLANE_A);
> > +	fifo_size = i915->display.get_fifo_size(i915, PLANE_A);
> > +	crtc = intel_get_crtc_for_plane(i915, PLANE_A);
> >   	if (intel_crtc_active(crtc)) {
> >   		const struct drm_display_mode *adjusted_mode =
> >   			&crtc->config->base.adjusted_mode;
> > @@ -2295,7 +2295,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   			crtc->base.primary->state->fb;
> >   		int cpp;
> > -		if (IS_GEN(dev_priv, 2))
> > +		if (IS_GEN(i915, 2))
> >   			cpp = 4;
> >   		else
> >   			cpp = fb->format->cpp[0];
> > @@ -2310,11 +2310,11 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   			planea_wm = wm_info->max_wm;
> >   	}
> > -	if (IS_GEN(dev_priv, 2))
> > +	if (IS_GEN(i915, 2))
> >   		wm_info = &i830_bc_wm_info;
> > -	fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_B);
> > -	crtc = intel_get_crtc_for_plane(dev_priv, PLANE_B);
> > +	fifo_size = i915->display.get_fifo_size(i915, PLANE_B);
> > +	crtc = intel_get_crtc_for_plane(i915, PLANE_B);
> >   	if (intel_crtc_active(crtc)) {
> >   		const struct drm_display_mode *adjusted_mode =
> >   			&crtc->config->base.adjusted_mode;
> > @@ -2322,7 +2322,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   			crtc->base.primary->state->fb;
> >   		int cpp;
> > -		if (IS_GEN(dev_priv, 2))
> > +		if (IS_GEN(i915, 2))
> >   			cpp = 4;
> >   		else
> >   			cpp = fb->format->cpp[0];
> > @@ -2342,7 +2342,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   	DRM_DEBUG_KMS("FIFO watermarks - A: %d, B: %d\n", planea_wm, planeb_wm);
> > -	if (IS_I915GM(dev_priv) && enabled) {
> > +	if (IS_I915GM(i915) && enabled) {
> >   		struct drm_i915_gem_object *obj;
> >   		obj = intel_fb_obj(enabled->base.primary->state->fb);
> > @@ -2358,10 +2358,10 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   	cwm = 2;
> >   	/* Play safe and disable self-refresh before adjusting watermarks. */
> > -	intel_set_memory_cxsr(dev_priv, false);
> > +	intel_set_memory_cxsr(i915, false);
> >   	/* Calc sr entries for one plane configs */
> > -	if (HAS_FW_BLC(dev_priv) && enabled) {
> > +	if (HAS_FW_BLC(i915) && enabled) {
> >   		/* self-refresh has much higher latency */
> >   		static const int sr_latency_ns = 6000;
> >   		const struct drm_display_mode *adjusted_mode =
> > @@ -2374,7 +2374,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   		int cpp;
> >   		int entries;
> > -		if (IS_I915GM(dev_priv) || IS_I945GM(dev_priv))
> > +		if (IS_I915GM(i915) || IS_I945GM(i915))
> >   			cpp = 4;
> >   		else
> >   			cpp = fb->format->cpp[0];
> > @@ -2387,7 +2387,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   		if (srwm < 0)
> >   			srwm = 1;
> > -		if (IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> > +		if (IS_I945G(i915) || IS_I945GM(i915))
> >   			I915_WRITE(FW_BLC_SELF,
> >   				   FW_BLC_SELF_FIFO_MASK | (srwm & 0xff));
> >   		else
> > @@ -2408,25 +2408,25 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc)
> >   	I915_WRITE(FW_BLC2, fwater_hi);
> >   	if (enabled)
> > -		intel_set_memory_cxsr(dev_priv, true);
> > +		intel_set_memory_cxsr(i915, true);
> >   }
> >   static void i845_update_wm(struct intel_crtc *unused_crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev);
> >   	struct intel_crtc *crtc;
> >   	const struct drm_display_mode *adjusted_mode;
> >   	u32 fwater_lo;
> >   	int planea_wm;
> > -	crtc = single_enabled_crtc(dev_priv);
> > +	crtc = single_enabled_crtc(i915);
> >   	if (crtc == NULL)
> >   		return;
> >   	adjusted_mode = &crtc->config->base.adjusted_mode;
> >   	planea_wm = intel_calculate_wm(adjusted_mode->crtc_clock,
> >   				       &i845_wm_info,
> > -				       dev_priv->display.get_fifo_size(dev_priv, PLANE_A),
> > +				       i915->display.get_fifo_size(i915, PLANE_A),
> >   				       4, pessimal_latency_ns);
> >   	fwater_lo = I915_READ(FW_BLC) & ~0xfff;
> >   	fwater_lo |= (3<<8) | planea_wm;
> > @@ -2586,24 +2586,24 @@ static u32 ilk_compute_fbc_wm(const struct intel_crtc_state *cstate,
> >   }
> >   static unsigned int
> > -ilk_display_fifo_size(const struct drm_i915_private *dev_priv)
> > +ilk_display_fifo_size(const struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		return 3072;
> > -	else if (INTEL_GEN(dev_priv) >= 7)
> > +	else if (INTEL_GEN(i915) >= 7)
> >   		return 768;
> >   	else
> >   		return 512;
> >   }
> >   static unsigned int
> > -ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv,
> > +ilk_plane_wm_reg_max(const struct drm_i915_private *i915,
> >   		     int level, bool is_sprite)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		/* BDW primary/sprite plane watermarks */
> >   		return level == 0 ? 255 : 2047;
> > -	else if (INTEL_GEN(dev_priv) >= 7)
> > +	else if (INTEL_GEN(i915) >= 7)
> >   		/* IVB/HSW primary/sprite plane watermarks */
> >   		return level == 0 ? 127 : 1023;
> >   	else if (!is_sprite)
> > @@ -2615,30 +2615,30 @@ ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv,
> >   }
> >   static unsigned int
> > -ilk_cursor_wm_reg_max(const struct drm_i915_private *dev_priv, int level)
> > +ilk_cursor_wm_reg_max(const struct drm_i915_private *i915, int level)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 7)
> > +	if (INTEL_GEN(i915) >= 7)
> >   		return level == 0 ? 63 : 255;
> >   	else
> >   		return level == 0 ? 31 : 63;
> >   }
> > -static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *dev_priv)
> > +static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *i915)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		return 31;
> >   	else
> >   		return 15;
> >   }
> >   /* Calculate the maximum primary/sprite plane watermark */
> > -static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
> > +static unsigned int ilk_plane_wm_max(const struct drm_i915_private *i915,
> >   				     int level,
> >   				     const struct intel_wm_config *config,
> >   				     enum intel_ddb_partitioning ddb_partitioning,
> >   				     bool is_sprite)
> >   {
> > -	unsigned int fifo_size = ilk_display_fifo_size(dev_priv);
> > +	unsigned int fifo_size = ilk_display_fifo_size(i915);
> >   	/* if sprites aren't enabled, sprites get nothing */
> >   	if (is_sprite && !config->sprites_enabled)
> > @@ -2646,14 +2646,14 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
> >   	/* HSW allows LP1+ watermarks even with multiple pipes */
> >   	if (level == 0 || config->num_pipes_active > 1) {
> > -		fifo_size /= INTEL_INFO(dev_priv)->num_pipes;
> > +		fifo_size /= INTEL_INFO(i915)->num_pipes;
> >   		/*
> >   		 * For some reason the non self refresh
> >   		 * FIFO size is only half of the self
> >   		 * refresh FIFO size on ILK/SNB.
> >   		 */
> > -		if (INTEL_GEN(dev_priv) <= 6)
> > +		if (INTEL_GEN(i915) <= 6)
> >   			fifo_size /= 2;
> >   	}
> > @@ -2669,11 +2669,11 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv,
> >   	}
> >   	/* clamp to max that the registers can hold */
> > -	return min(fifo_size, ilk_plane_wm_reg_max(dev_priv, level, is_sprite));
> > +	return min(fifo_size, ilk_plane_wm_reg_max(i915, level, is_sprite));
> >   }
> >   /* Calculate the maximum cursor plane watermark */
> > -static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv,
> > +static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *i915,
> >   				      int level,
> >   				      const struct intel_wm_config *config)
> >   {
> > @@ -2682,29 +2682,29 @@ static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv,
> >   		return 64;
> >   	/* otherwise just report max that registers can hold */
> > -	return ilk_cursor_wm_reg_max(dev_priv, level);
> > +	return ilk_cursor_wm_reg_max(i915, level);
> >   }
> > -static void ilk_compute_wm_maximums(const struct drm_i915_private *dev_priv,
> > +static void ilk_compute_wm_maximums(const struct drm_i915_private *i915,
> >   				    int level,
> >   				    const struct intel_wm_config *config,
> >   				    enum intel_ddb_partitioning ddb_partitioning,
> >   				    struct ilk_wm_maximums *max)
> >   {
> > -	max->pri = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, false);
> > -	max->spr = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, true);
> > -	max->cur = ilk_cursor_wm_max(dev_priv, level, config);
> > -	max->fbc = ilk_fbc_wm_reg_max(dev_priv);
> > +	max->pri = ilk_plane_wm_max(i915, level, config, ddb_partitioning, false);
> > +	max->spr = ilk_plane_wm_max(i915, level, config, ddb_partitioning, true);
> > +	max->cur = ilk_cursor_wm_max(i915, level, config);
> > +	max->fbc = ilk_fbc_wm_reg_max(i915);
> >   }
> > -static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *dev_priv,
> > +static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *i915,
> >   					int level,
> >   					struct ilk_wm_maximums *max)
> >   {
> > -	max->pri = ilk_plane_wm_reg_max(dev_priv, level, false);
> > -	max->spr = ilk_plane_wm_reg_max(dev_priv, level, true);
> > -	max->cur = ilk_cursor_wm_reg_max(dev_priv, level);
> > -	max->fbc = ilk_fbc_wm_reg_max(dev_priv);
> > +	max->pri = ilk_plane_wm_reg_max(i915, level, false);
> > +	max->spr = ilk_plane_wm_reg_max(i915, level, true);
> > +	max->cur = ilk_cursor_wm_reg_max(i915, level);
> > +	max->fbc = ilk_fbc_wm_reg_max(i915);
> >   }
> >   static bool ilk_validate_wm_level(int level,
> > @@ -2748,7 +2748,7 @@ static bool ilk_validate_wm_level(int level,
> >   	return ret;
> >   }
> > -static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv,
> > +static void ilk_compute_wm_level(const struct drm_i915_private *i915,
> >   				 const struct intel_crtc *intel_crtc,
> >   				 int level,
> >   				 struct intel_crtc_state *cstate,
> > @@ -2757,9 +2757,9 @@ static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv,
> >   				 const struct intel_plane_state *curstate,
> >   				 struct intel_wm_level *result)
> >   {
> > -	u16 pri_latency = dev_priv->wm.pri_latency[level];
> > -	u16 spr_latency = dev_priv->wm.spr_latency[level];
> > -	u16 cur_latency = dev_priv->wm.cur_latency[level];
> > +	u16 pri_latency = i915->wm.pri_latency[level];
> > +	u16 spr_latency = i915->wm.spr_latency[level];
> > +	u16 cur_latency = i915->wm.cur_latency[level];
> >   	/* WM1+ latency values stored in 0.5us units */
> >   	if (level > 0) {
> > @@ -2811,19 +2811,19 @@ hsw_compute_linetime_wm(const struct intel_crtc_state *cstate)
> >   	       PIPE_WM_LINETIME_TIME(linetime);
> >   }
> > -static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> > +static void intel_read_wm_latency(struct drm_i915_private *i915,
> >   				  u16 wm[8])
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		u32 val;
> >   		int ret, i;
> > -		int level, max_level = ilk_wm_max_level(dev_priv);
> > +		int level, max_level = ilk_wm_max_level(i915);
> >   		/* read the first set of memory latencies[0:3] */
> >   		val = 0; /* data0 to be programmed to 0 for first set */
> > -		ret = sandybridge_pcode_read(dev_priv,
> > +		ret = sandybridge_pcode_read(i915,
> >   					     GEN9_PCODE_READ_MEM_LATENCY,
> >   					     &val, NULL);
> > @@ -2842,7 +2842,7 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> >   		/* read the second set of memory latencies[4:7] */
> >   		val = 1; /* data0 to be programmed to 1 for second set */
> > -		ret = sandybridge_pcode_read(dev_priv,
> > +		ret = sandybridge_pcode_read(i915,
> >   					     GEN9_PCODE_READ_MEM_LATENCY,
> >   					     &val, NULL);
> >   		if (ret) {
> > @@ -2893,10 +2893,10 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> >   		 * any underrun. If not able to get Dimm info assume 16GB dimm
> >   		 * to avoid any underrun.
> >   		 */
> > -		if (dev_priv->dram_info.is_16gb_dimm)
> > +		if (i915->dram_info.is_16gb_dimm)
> >   			wm[0] += 1;
> > -	} else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	} else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		u64 sskpd = intel_uncore_read64(uncore, MCH_SSKPD);
> >   		wm[0] = (sskpd >> 56) & 0xFF;
> > @@ -2906,14 +2906,14 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> >   		wm[2] = (sskpd >> 12) & 0xFF;
> >   		wm[3] = (sskpd >> 20) & 0x1FF;
> >   		wm[4] = (sskpd >> 32) & 0x1FF;
> > -	} else if (INTEL_GEN(dev_priv) >= 6) {
> > +	} else if (INTEL_GEN(i915) >= 6) {
> >   		u32 sskpd = intel_uncore_read(uncore, MCH_SSKPD);
> >   		wm[0] = (sskpd >> SSKPD_WM0_SHIFT) & SSKPD_WM_MASK;
> >   		wm[1] = (sskpd >> SSKPD_WM1_SHIFT) & SSKPD_WM_MASK;
> >   		wm[2] = (sskpd >> SSKPD_WM2_SHIFT) & SSKPD_WM_MASK;
> >   		wm[3] = (sskpd >> SSKPD_WM3_SHIFT) & SSKPD_WM_MASK;
> > -	} else if (INTEL_GEN(dev_priv) >= 5) {
> > +	} else if (INTEL_GEN(i915) >= 5) {
> >   		u32 mltr = intel_uncore_read(uncore, MLTR_ILK);
> >   		/* ILK primary LP0 latency is 700 ns */
> > @@ -2921,44 +2921,44 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv,
> >   		wm[1] = (mltr >> MLTR_WM1_SHIFT) & ILK_SRLT_MASK;
> >   		wm[2] = (mltr >> MLTR_WM2_SHIFT) & ILK_SRLT_MASK;
> >   	} else {
> > -		MISSING_CASE(INTEL_DEVID(dev_priv));
> > +		MISSING_CASE(INTEL_DEVID(i915));
> >   	}
> >   }
> > -static void intel_fixup_spr_wm_latency(struct drm_i915_private *dev_priv,
> > +static void intel_fixup_spr_wm_latency(struct drm_i915_private *i915,
> >   				       u16 wm[5])
> >   {
> >   	/* ILK sprite LP0 latency is 1300 ns */
> > -	if (IS_GEN(dev_priv, 5))
> > +	if (IS_GEN(i915, 5))
> >   		wm[0] = 13;
> >   }
> > -static void intel_fixup_cur_wm_latency(struct drm_i915_private *dev_priv,
> > +static void intel_fixup_cur_wm_latency(struct drm_i915_private *i915,
> >   				       u16 wm[5])
> >   {
> >   	/* ILK cursor LP0 latency is 1300 ns */
> > -	if (IS_GEN(dev_priv, 5))
> > +	if (IS_GEN(i915, 5))
> >   		wm[0] = 13;
> >   }
> > -int ilk_wm_max_level(const struct drm_i915_private *dev_priv)
> > +int ilk_wm_max_level(const struct drm_i915_private *i915)
> >   {
> >   	/* how many WM levels are we expecting */
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return 7;
> > -	else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		return 4;
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > +	else if (INTEL_GEN(i915) >= 6)
> >   		return 3;
> >   	else
> >   		return 2;
> >   }
> > -static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> > +static void intel_print_wm_latency(struct drm_i915_private *i915,
> >   				   const char *name,
> >   				   const u16 wm[8])
> >   {
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	for (level = 0; level <= max_level; level++) {
> >   		unsigned int latency = wm[level];
> > @@ -2973,7 +2973,7 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> >   		 * - latencies are in us on gen9.
> >   		 * - before then, WM1+ latency values are in 0.5us units
> >   		 */
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			latency *= 10;
> >   		else if (level > 0)
> >   			latency *= 5;
> > @@ -2984,10 +2984,10 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv,
> > +static bool ilk_increase_wm_latency(struct drm_i915_private *i915,
> >   				    u16 wm[5], u16 min)
> >   {
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	if (wm[0] >= min)
> >   		return false;
> > @@ -2999,7 +2999,7 @@ static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv)
> > +static void snb_wm_latency_quirk(struct drm_i915_private *i915)
> >   {
> >   	bool changed;
> > @@ -3007,20 +3007,20 @@ static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv)
> >   	 * The BIOS provided WM memory latency values are often
> >   	 * inadequate for high resolution displays. Adjust them.
> >   	 */
> > -	changed = ilk_increase_wm_latency(dev_priv, dev_priv->wm.pri_latency, 12) |
> > -		ilk_increase_wm_latency(dev_priv, dev_priv->wm.spr_latency, 12) |
> > -		ilk_increase_wm_latency(dev_priv, dev_priv->wm.cur_latency, 12);
> > +	changed = ilk_increase_wm_latency(i915, i915->wm.pri_latency, 12) |
> > +		ilk_increase_wm_latency(i915, i915->wm.spr_latency, 12) |
> > +		ilk_increase_wm_latency(i915, i915->wm.cur_latency, 12);
> >   	if (!changed)
> >   		return;
> >   	DRM_DEBUG_KMS("WM latency values increased to avoid potential underruns\n");
> > -	intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> > -	intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> > -	intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> > +	intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> > +	intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> > +	intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
> >   }
> > -static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv)
> > +static void snb_wm_lp3_irq_quirk(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * On some SNB machines (Thinkpad X220 Tablet at least)
> > @@ -3033,50 +3033,50 @@ static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv)
> >   	 * interrupts only. To play it safe we disable LP3
> >   	 * watermarks entirely.
> >   	 */
> > -	if (dev_priv->wm.pri_latency[3] == 0 &&
> > -	    dev_priv->wm.spr_latency[3] == 0 &&
> > -	    dev_priv->wm.cur_latency[3] == 0)
> > +	if (i915->wm.pri_latency[3] == 0 &&
> > +	    i915->wm.spr_latency[3] == 0 &&
> > +	    i915->wm.cur_latency[3] == 0)
> >   		return;
> > -	dev_priv->wm.pri_latency[3] = 0;
> > -	dev_priv->wm.spr_latency[3] = 0;
> > -	dev_priv->wm.cur_latency[3] = 0;
> > +	i915->wm.pri_latency[3] = 0;
> > +	i915->wm.spr_latency[3] = 0;
> > +	i915->wm.cur_latency[3] = 0;
> >   	DRM_DEBUG_KMS("LP3 watermarks disabled due to potential for lost interrupts\n");
> > -	intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> > -	intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> > -	intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> > +	intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> > +	intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> > +	intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
> >   }
> > -static void ilk_setup_wm_latency(struct drm_i915_private *dev_priv)
> > +static void ilk_setup_wm_latency(struct drm_i915_private *i915)
> >   {
> > -	intel_read_wm_latency(dev_priv, dev_priv->wm.pri_latency);
> > +	intel_read_wm_latency(i915, i915->wm.pri_latency);
> > -	memcpy(dev_priv->wm.spr_latency, dev_priv->wm.pri_latency,
> > -	       sizeof(dev_priv->wm.pri_latency));
> > -	memcpy(dev_priv->wm.cur_latency, dev_priv->wm.pri_latency,
> > -	       sizeof(dev_priv->wm.pri_latency));
> > +	memcpy(i915->wm.spr_latency, i915->wm.pri_latency,
> > +	       sizeof(i915->wm.pri_latency));
> > +	memcpy(i915->wm.cur_latency, i915->wm.pri_latency,
> > +	       sizeof(i915->wm.pri_latency));
> > -	intel_fixup_spr_wm_latency(dev_priv, dev_priv->wm.spr_latency);
> > -	intel_fixup_cur_wm_latency(dev_priv, dev_priv->wm.cur_latency);
> > +	intel_fixup_spr_wm_latency(i915, i915->wm.spr_latency);
> > +	intel_fixup_cur_wm_latency(i915, i915->wm.cur_latency);
> > -	intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
> > -	intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
> > -	intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
> > +	intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency);
> > +	intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency);
> > +	intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency);
> > -	if (IS_GEN(dev_priv, 6)) {
> > -		snb_wm_latency_quirk(dev_priv);
> > -		snb_wm_lp3_irq_quirk(dev_priv);
> > +	if (IS_GEN(i915, 6)) {
> > +		snb_wm_latency_quirk(i915);
> > +		snb_wm_lp3_irq_quirk(i915);
> >   	}
> >   }
> > -static void skl_setup_wm_latency(struct drm_i915_private *dev_priv)
> > +static void skl_setup_wm_latency(struct drm_i915_private *i915)
> >   {
> > -	intel_read_wm_latency(dev_priv, dev_priv->wm.skl_latency);
> > -	intel_print_wm_latency(dev_priv, "Gen9 Plane", dev_priv->wm.skl_latency);
> > +	intel_read_wm_latency(i915, i915->wm.skl_latency);
> > +	intel_print_wm_latency(i915, "Gen9 Plane", i915->wm.skl_latency);
> >   }
> > -static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv,
> > +static bool ilk_validate_pipe_wm(const struct drm_i915_private *i915,
> >   				 struct intel_pipe_wm *pipe_wm)
> >   {
> >   	/* LP0 watermark maximums depend on this pipe alone */
> > @@ -3088,7 +3088,7 @@ static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv,
> >   	struct ilk_wm_maximums max;
> >   	/* LP0 watermarks always use 1/2 DDB partitioning */
> > -	ilk_compute_wm_maximums(dev_priv, 0, &config, INTEL_DDB_PART_1_2, &max);
> > +	ilk_compute_wm_maximums(i915, 0, &config, INTEL_DDB_PART_1_2, &max);
> >   	/* At least LP0 must be valid */
> >   	if (!ilk_validate_wm_level(0, &max, &pipe_wm->wm[0])) {
> > @@ -3106,13 +3106,13 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> >   	struct intel_pipe_wm *pipe_wm;
> >   	struct drm_device *dev = state->dev;
> > -	const struct drm_i915_private *dev_priv = to_i915(dev);
> > +	const struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_plane *plane;
> >   	const struct drm_plane_state *plane_state;
> >   	const struct intel_plane_state *pristate = NULL;
> >   	const struct intel_plane_state *sprstate = NULL;
> >   	const struct intel_plane_state *curstate = NULL;
> > -	int level, max_level = ilk_wm_max_level(dev_priv), usable_level;
> > +	int level, max_level = ilk_wm_max_level(i915), usable_level;
> >   	struct ilk_wm_maximums max;
> >   	pipe_wm = &cstate->wm.ilk.optimal;
> > @@ -3139,7 +3139,7 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> >   	usable_level = max_level;
> >   	/* ILK/SNB: LP2+ watermarks only w/o sprites */
> > -	if (INTEL_GEN(dev_priv) <= 6 && pipe_wm->sprites_enabled)
> > +	if (INTEL_GEN(i915) <= 6 && pipe_wm->sprites_enabled)
> >   		usable_level = 1;
> >   	/* ILK/SNB/IVB: LP1+ watermarks only w/o scaling */
> > @@ -3147,21 +3147,21 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> >   		usable_level = 0;
> >   	memset(&pipe_wm->wm, 0, sizeof(pipe_wm->wm));
> > -	ilk_compute_wm_level(dev_priv, intel_crtc, 0, cstate,
> > +	ilk_compute_wm_level(i915, intel_crtc, 0, cstate,
> >   			     pristate, sprstate, curstate, &pipe_wm->wm[0]);
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		pipe_wm->linetime = hsw_compute_linetime_wm(cstate);
> > -	if (!ilk_validate_pipe_wm(dev_priv, pipe_wm))
> > +	if (!ilk_validate_pipe_wm(i915, pipe_wm))
> >   		return -EINVAL;
> > -	ilk_compute_wm_reg_maximums(dev_priv, 1, &max);
> > +	ilk_compute_wm_reg_maximums(i915, 1, &max);
> >   	for (level = 1; level <= usable_level; level++) {
> >   		struct intel_wm_level *wm = &pipe_wm->wm[level];
> > -		ilk_compute_wm_level(dev_priv, intel_crtc, level, cstate,
> > +		ilk_compute_wm_level(i915, intel_crtc, level, cstate,
> >   				     pristate, sprstate, curstate, wm);
> >   		/*
> > @@ -3186,14 +3186,14 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate)
> >   static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> >   {
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(newstate->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	struct intel_pipe_wm *a = &newstate->wm.ilk.intermediate;
> >   	struct intel_atomic_state *intel_state =
> >   		to_intel_atomic_state(newstate->base.state);
> >   	const struct intel_crtc_state *oldstate =
> >   		intel_atomic_get_old_crtc_state(intel_state, intel_crtc);
> >   	const struct intel_pipe_wm *b = &oldstate->wm.ilk.optimal;
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	/*
> >   	 * Start with the final, target watermarks, then combine with the
> > @@ -3226,7 +3226,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> >   	 * there's no safe way to transition from the old state to
> >   	 * the new state, so we need to fail the atomic transaction.
> >   	 */
> > -	if (!ilk_validate_pipe_wm(dev_priv, a))
> > +	if (!ilk_validate_pipe_wm(i915, a))
> >   		return -EINVAL;
> >   	/*
> > @@ -3242,7 +3242,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate)
> >   /*
> >    * Merge the watermarks from all active pipes for a specific level.
> >    */
> > -static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
> > +static void ilk_merge_wm_level(struct drm_i915_private *i915,
> >   			       int level,
> >   			       struct intel_wm_level *ret_wm)
> >   {
> > @@ -3250,7 +3250,7 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
> >   	ret_wm->enable = true;
> > -	for_each_intel_crtc(&dev_priv->drm, intel_crtc) {
> > +	for_each_intel_crtc(&i915->drm, intel_crtc) {
> >   		const struct intel_pipe_wm *active = &intel_crtc->wm.active.ilk;
> >   		const struct intel_wm_level *wm = &active->wm[level];
> > @@ -3275,27 +3275,27 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv,
> >   /*
> >    * Merge all low power watermarks for all active pipes.
> >    */
> > -static void ilk_wm_merge(struct drm_i915_private *dev_priv,
> > +static void ilk_wm_merge(struct drm_i915_private *i915,
> >   			 const struct intel_wm_config *config,
> >   			 const struct ilk_wm_maximums *max,
> >   			 struct intel_pipe_wm *merged)
> >   {
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	int last_enabled_level = max_level;
> >   	/* ILK/SNB/IVB: LP1+ watermarks only w/ single pipe */
> > -	if ((INTEL_GEN(dev_priv) <= 6 || IS_IVYBRIDGE(dev_priv)) &&
> > +	if ((INTEL_GEN(i915) <= 6 || IS_IVYBRIDGE(i915)) &&
> >   	    config->num_pipes_active > 1)
> >   		last_enabled_level = 0;
> >   	/* ILK: FBC WM must be disabled always */
> > -	merged->fbc_wm_enabled = INTEL_GEN(dev_priv) >= 6;
> > +	merged->fbc_wm_enabled = INTEL_GEN(i915) >= 6;
> >   	/* merge each WM1+ level */
> >   	for (level = 1; level <= max_level; level++) {
> >   		struct intel_wm_level *wm = &merged->wm[level];
> > -		ilk_merge_wm_level(dev_priv, level, wm);
> > +		ilk_merge_wm_level(i915, level, wm);
> >   		if (level > last_enabled_level)
> >   			wm->enable = false;
> > @@ -3320,8 +3320,8 @@ static void ilk_wm_merge(struct drm_i915_private *dev_priv,
> >   	 * What we should check here is whether FBC can be
> >   	 * enabled sometime later.
> >   	 */
> > -	if (IS_GEN(dev_priv, 5) && !merged->fbc_wm_enabled &&
> > -	    intel_fbc_is_active(dev_priv)) {
> > +	if (IS_GEN(i915, 5) && !merged->fbc_wm_enabled &&
> > +	    intel_fbc_is_active(i915)) {
> >   		for (level = 2; level <= max_level; level++) {
> >   			struct intel_wm_level *wm = &merged->wm[level];
> > @@ -3337,16 +3337,16 @@ static int ilk_wm_lp_to_level(int wm_lp, const struct intel_pipe_wm *pipe_wm)
> >   }
> >   /* The value we need to program into the WM_LPx latency field */
> > -static unsigned int ilk_wm_lp_latency(struct drm_i915_private *dev_priv,
> > +static unsigned int ilk_wm_lp_latency(struct drm_i915_private *i915,
> >   				      int level)
> >   {
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		return 2 * level;
> >   	else
> > -		return dev_priv->wm.pri_latency[level];
> > +		return i915->wm.pri_latency[level];
> >   }
> > -static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> > +static void ilk_compute_wm_results(struct drm_i915_private *i915,
> >   				   const struct intel_pipe_wm *merged,
> >   				   enum intel_ddb_partitioning partitioning,
> >   				   struct ilk_wm_values *results)
> > @@ -3370,14 +3370,14 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> >   		 * disabled. Doing otherwise could cause underruns.
> >   		 */
> >   		results->wm_lp[wm_lp - 1] =
> > -			(ilk_wm_lp_latency(dev_priv, level) << WM1_LP_LATENCY_SHIFT) |
> > +			(ilk_wm_lp_latency(i915, level) << WM1_LP_LATENCY_SHIFT) |
> >   			(r->pri_val << WM1_LP_SR_SHIFT) |
> >   			r->cur_val;
> >   		if (r->enable)
> >   			results->wm_lp[wm_lp - 1] |= WM1_LP_SR_EN;
> > -		if (INTEL_GEN(dev_priv) >= 8)
> > +		if (INTEL_GEN(i915) >= 8)
> >   			results->wm_lp[wm_lp - 1] |=
> >   				r->fbc_val << WM1_LP_FBC_SHIFT_BDW;
> >   		else
> > @@ -3388,7 +3388,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> >   		 * Always set WM1S_LP_EN when spr_val != 0, even if the
> >   		 * level is disabled. Doing otherwise could cause underruns.
> >   		 */
> > -		if (INTEL_GEN(dev_priv) <= 6 && r->spr_val) {
> > +		if (INTEL_GEN(i915) <= 6 && r->spr_val) {
> >   			WARN_ON(wm_lp != 1);
> >   			results->wm_lp_spr[wm_lp - 1] = WM1S_LP_EN | r->spr_val;
> >   		} else
> > @@ -3396,7 +3396,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> >   	}
> >   	/* LP0 register values */
> > -	for_each_intel_crtc(&dev_priv->drm, intel_crtc) {
> > +	for_each_intel_crtc(&i915->drm, intel_crtc) {
> >   		enum pipe pipe = intel_crtc->pipe;
> >   		const struct intel_wm_level *r =
> >   			&intel_crtc->wm.active.ilk.wm[0];
> > @@ -3416,11 +3416,11 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
> >   /* Find the result with the highest level enabled. Check for enable_fbc_wm in
> >    * case both are at the same level. Prefer r1 in case they're the same. */
> >   static struct intel_pipe_wm *
> > -ilk_find_best_result(struct drm_i915_private *dev_priv,
> > +ilk_find_best_result(struct drm_i915_private *i915,
> >   		     struct intel_pipe_wm *r1,
> >   		     struct intel_pipe_wm *r2)
> >   {
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	int level1 = 0, level2 = 0;
> >   	for (level = 1; level <= max_level; level++) {
> > @@ -3450,7 +3450,7 @@ ilk_find_best_result(struct drm_i915_private *dev_priv,
> >   #define WM_DIRTY_FBC (1 << 24)
> >   #define WM_DIRTY_DDB (1 << 25)
> > -static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> > +static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *i915,
> >   					 const struct ilk_wm_values *old,
> >   					 const struct ilk_wm_values *new)
> >   {
> > @@ -3458,7 +3458,7 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> >   	enum pipe pipe;
> >   	int wm_lp;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		if (old->wm_linetime[pipe] != new->wm_linetime[pipe]) {
> >   			dirty |= WM_DIRTY_LINETIME(pipe);
> >   			/* Must disable LP1+ watermarks too */
> > @@ -3502,10 +3502,10 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv,
> >   	return dirty;
> >   }
> > -static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv,
> > +static bool _ilk_disable_lp_wm(struct drm_i915_private *i915,
> >   			       unsigned int dirty)
> >   {
> > -	struct ilk_wm_values *previous = &dev_priv->wm.hw;
> > +	struct ilk_wm_values *previous = &i915->wm.hw;
> >   	bool changed = false;
> >   	if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] & WM1_LP_SR_EN) {
> > @@ -3536,18 +3536,18 @@ static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv,
> >    * The spec says we shouldn't write when we don't need, because every write
> >    * causes WMs to be re-evaluated, expending some power.
> >    */
> > -static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> > +static void ilk_write_wm_values(struct drm_i915_private *i915,
> >   				struct ilk_wm_values *results)
> >   {
> > -	struct ilk_wm_values *previous = &dev_priv->wm.hw;
> > +	struct ilk_wm_values *previous = &i915->wm.hw;
> >   	unsigned int dirty;
> >   	u32 val;
> > -	dirty = ilk_compute_wm_dirty(dev_priv, previous, results);
> > +	dirty = ilk_compute_wm_dirty(i915, previous, results);
> >   	if (!dirty)
> >   		return;
> > -	_ilk_disable_lp_wm(dev_priv, dirty);
> > +	_ilk_disable_lp_wm(i915, dirty);
> >   	if (dirty & WM_DIRTY_PIPE(PIPE_A))
> >   		I915_WRITE(WM0_PIPEA_ILK, results->wm_pipe[0]);
> > @@ -3564,7 +3564,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> >   		I915_WRITE(PIPE_WM_LINETIME(PIPE_C), results->wm_linetime[2]);
> >   	if (dirty & WM_DIRTY_DDB) {
> > -		if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +		if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   			val = I915_READ(WM_MISC);
> >   			if (results->partitioning == INTEL_DDB_PART_1_2)
> >   				val &= ~WM_MISC_DATA_PARTITION_5_6;
> > @@ -3594,7 +3594,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> >   	    previous->wm_lp_spr[0] != results->wm_lp_spr[0])
> >   		I915_WRITE(WM1S_LP_ILK, results->wm_lp_spr[0]);
> > -	if (INTEL_GEN(dev_priv) >= 7) {
> > +	if (INTEL_GEN(i915) >= 7) {
> >   		if (dirty & WM_DIRTY_LP(2) && previous->wm_lp_spr[1] != results->wm_lp_spr[1])
> >   			I915_WRITE(WM2S_LP_IVB, results->wm_lp_spr[1]);
> >   		if (dirty & WM_DIRTY_LP(3) && previous->wm_lp_spr[2] != results->wm_lp_spr[2])
> > @@ -3608,17 +3608,17 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
> >   	if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] != results->wm_lp[2])
> >   		I915_WRITE(WM3_LP_ILK, results->wm_lp[2]);
> > -	dev_priv->wm.hw = *results;
> > +	i915->wm.hw = *results;
> >   }
> >   bool ilk_disable_lp_wm(struct drm_device *dev)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > -	return _ilk_disable_lp_wm(dev_priv, WM_DIRTY_LP_ALL);
> > +	return _ilk_disable_lp_wm(i915, WM_DIRTY_LP_ALL);
> >   }
> > -static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> > +static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *i915)
> >   {
> >   	u8 enabled_slices;
> > @@ -3626,7 +3626,7 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> >   	enabled_slices = 1;
> >   	/* Gen prior to GEN11 have only one DBuf slice */
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return enabled_slices;
> >   	/*
> > @@ -3644,16 +3644,16 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv)
> >    * FIXME: We still don't have the proper code detect if we need to apply the WA,
> >    * so assume we'll always need it in order to avoid underruns.
> >    */
> > -static bool skl_needs_memory_bw_wa(struct drm_i915_private *dev_priv)
> > +static bool skl_needs_memory_bw_wa(struct drm_i915_private *i915)
> >   {
> > -	return IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv);
> > +	return IS_GEN9_BC(i915) || IS_BROXTON(i915);
> >   }
> >   static bool
> > -intel_has_sagv(struct drm_i915_private *dev_priv)
> > +intel_has_sagv(struct drm_i915_private *i915)
> >   {
> > -	return (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) &&
> > -		dev_priv->sagv_status != I915_SAGV_NOT_CONTROLLED;
> > +	return (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) &&
> > +		i915->sagv_status != I915_SAGV_NOT_CONTROLLED;
> >   }
> >   /*
> > @@ -3668,18 +3668,18 @@ intel_has_sagv(struct drm_i915_private *dev_priv)
> >    *  - We're not using an interlaced display configuration
> >    */
> >   int
> > -intel_enable_sagv(struct drm_i915_private *dev_priv)
> > +intel_enable_sagv(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	if (!intel_has_sagv(dev_priv))
> > +	if (!intel_has_sagv(i915))
> >   		return 0;
> > -	if (dev_priv->sagv_status == I915_SAGV_ENABLED)
> > +	if (i915->sagv_status == I915_SAGV_ENABLED)
> >   		return 0;
> >   	DRM_DEBUG_KMS("Enabling SAGV\n");
> > -	ret = sandybridge_pcode_write(dev_priv, GEN9_PCODE_SAGV_CONTROL,
> > +	ret = sandybridge_pcode_write(i915, GEN9_PCODE_SAGV_CONTROL,
> >   				      GEN9_SAGV_ENABLE);
> >   	/* We don't need to wait for SAGV when enabling */
> > @@ -3688,33 +3688,33 @@ intel_enable_sagv(struct drm_i915_private *dev_priv)
> >   	 * Some skl systems, pre-release machines in particular,
> >   	 * don't actually have SAGV.
> >   	 */
> > -	if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) {
> > +	if (IS_SKYLAKE(i915) && ret == -ENXIO) {
> >   		DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n");
> > -		dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED;
> > +		i915->sagv_status = I915_SAGV_NOT_CONTROLLED;
> >   		return 0;
> >   	} else if (ret < 0) {
> >   		DRM_ERROR("Failed to enable SAGV\n");
> >   		return ret;
> >   	}
> > -	dev_priv->sagv_status = I915_SAGV_ENABLED;
> > +	i915->sagv_status = I915_SAGV_ENABLED;
> >   	return 0;
> >   }
> >   int
> > -intel_disable_sagv(struct drm_i915_private *dev_priv)
> > +intel_disable_sagv(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> > -	if (!intel_has_sagv(dev_priv))
> > +	if (!intel_has_sagv(i915))
> >   		return 0;
> > -	if (dev_priv->sagv_status == I915_SAGV_DISABLED)
> > +	if (i915->sagv_status == I915_SAGV_DISABLED)
> >   		return 0;
> >   	DRM_DEBUG_KMS("Disabling SAGV\n");
> >   	/* bspec says to keep retrying for at least 1 ms */
> > -	ret = skl_pcode_request(dev_priv, GEN9_PCODE_SAGV_CONTROL,
> > +	ret = skl_pcode_request(i915, GEN9_PCODE_SAGV_CONTROL,
> >   				GEN9_SAGV_DISABLE,
> >   				GEN9_SAGV_IS_DISABLED, GEN9_SAGV_IS_DISABLED,
> >   				1);
> > @@ -3722,23 +3722,23 @@ intel_disable_sagv(struct drm_i915_private *dev_priv)
> >   	 * Some skl systems, pre-release machines in particular,
> >   	 * don't actually have SAGV.
> >   	 */
> > -	if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) {
> > +	if (IS_SKYLAKE(i915) && ret == -ENXIO) {
> >   		DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n");
> > -		dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED;
> > +		i915->sagv_status = I915_SAGV_NOT_CONTROLLED;
> >   		return 0;
> >   	} else if (ret < 0) {
> >   		DRM_ERROR("Failed to disable SAGV (%d)\n", ret);
> >   		return ret;
> >   	}
> > -	dev_priv->sagv_status = I915_SAGV_DISABLED;
> > +	i915->sagv_status = I915_SAGV_DISABLED;
> >   	return 0;
> >   }
> >   bool intel_can_enable_sagv(struct drm_atomic_state *state)
> >   {
> >   	struct drm_device *dev = state->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> >   	struct intel_crtc *crtc;
> >   	struct intel_plane *plane;
> > @@ -3747,12 +3747,12 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> >   	int level, latency;
> >   	int sagv_block_time_us;
> > -	if (!intel_has_sagv(dev_priv))
> > +	if (!intel_has_sagv(i915))
> >   		return false;
> > -	if (IS_GEN(dev_priv, 9))
> > +	if (IS_GEN(i915, 9))
> >   		sagv_block_time_us = 30;
> > -	else if (IS_GEN(dev_priv, 10))
> > +	else if (IS_GEN(i915, 10))
> >   		sagv_block_time_us = 20;
> >   	else
> >   		sagv_block_time_us = 10;
> > @@ -3772,7 +3772,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> >   	/* Since we're now guaranteed to only have one active CRTC... */
> >   	pipe = ffs(intel_state->active_crtcs) - 1;
> > -	crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	cstate = to_intel_crtc_state(crtc->base.state);
> >   	if (crtc->base.state->adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
> > @@ -3787,13 +3787,13 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> >   			continue;
> >   		/* Find the highest enabled wm level for this plane */
> > -		for (level = ilk_wm_max_level(dev_priv);
> > +		for (level = ilk_wm_max_level(i915);
> >   		     !wm->wm[level].plane_en; --level)
> >   		     { }
> > -		latency = dev_priv->wm.skl_latency[level];
> > +		latency = i915->wm.skl_latency[level];
> > -		if (skl_needs_memory_bw_wa(dev_priv) &&
> > +		if (skl_needs_memory_bw_wa(i915) &&
> >   		    plane->base.state->fb->modifier ==
> >   		    I915_FORMAT_MOD_X_TILED)
> >   			latency += 15;
> > @@ -3810,7 +3810,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state)
> >   	return true;
> >   }
> > -static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> > +static u16 intel_get_ddb_size(struct drm_i915_private *i915,
> >   			      const struct intel_crtc_state *cstate,
> >   			      const u64 total_data_rate,
> >   			      const int num_active,
> > @@ -3818,11 +3818,11 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> >   {
> >   	const struct drm_display_mode *adjusted_mode;
> >   	u64 total_data_bw;
> > -	u16 ddb_size = INTEL_INFO(dev_priv)->ddb_size;
> > +	u16 ddb_size = INTEL_INFO(i915)->ddb_size;
> >   	WARN_ON(ddb_size == 0);
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		return ddb_size - 4; /* 4 blocks for bypass path allocation */
> >   	adjusted_mode = &cstate->base.adjusted_mode;
> > @@ -3847,7 +3847,7 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv,
> > +skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *i915,
> >   				   const struct intel_crtc_state *cstate,
> >   				   const u64 total_data_rate,
> >   				   struct skl_ddb_allocation *ddb,
> > @@ -3867,16 +3867,16 @@ skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv,
> >   	if (WARN_ON(!state) || !cstate->base.active) {
> >   		alloc->start = 0;
> >   		alloc->end = 0;
> > -		*num_active = hweight32(dev_priv->active_crtcs);
> > +		*num_active = hweight32(i915->active_crtcs);
> >   		return;
> >   	}
> >   	if (intel_state->active_pipe_changes)
> >   		*num_active = hweight32(intel_state->active_crtcs);
> >   	else
> > -		*num_active = hweight32(dev_priv->active_crtcs);
> > +		*num_active = hweight32(i915->active_crtcs);
> > -	ddb_size = intel_get_ddb_size(dev_priv, cstate, total_data_rate,
> > +	ddb_size = intel_get_ddb_size(i915, cstate, total_data_rate,
> >   				      *num_active, ddb);
> >   	/*
> > @@ -3939,8 +3939,8 @@ static unsigned int
> >   skl_cursor_allocation(const struct intel_crtc_state *crtc_state,
> >   		      int num_active)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev);
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	struct skl_wm_level wm = {};
> >   	int ret, min_ddb_alloc = 0;
> >   	struct skl_wm_params wp;
> > @@ -3963,7 +3963,7 @@ skl_cursor_allocation(const struct intel_crtc_state *crtc_state,
> >   	return max(num_active == 1 ? 32 : 8, min_ddb_alloc);
> >   }
> > -static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv,
> > +static void skl_ddb_entry_init_from_hw(struct drm_i915_private *i915,
> >   				       struct skl_ddb_entry *entry, u32 reg)
> >   {
> > @@ -3975,7 +3975,7 @@ static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> > +skl_ddb_get_hw_plane_state(struct drm_i915_private *i915,
> >   			   const enum pipe pipe,
> >   			   const enum plane_id plane_id,
> >   			   struct skl_ddb_entry *ddb_y,
> > @@ -3987,7 +3987,7 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> >   	/* Cursor doesn't support NV12/planar, so no extra calculation needed */
> >   	if (plane_id == PLANE_CURSOR) {
> >   		val = I915_READ(CUR_BUF_CFG(pipe));
> > -		skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> > +		skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> >   		return;
> >   	}
> > @@ -3999,9 +3999,9 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> >   					      val & PLANE_CTL_ORDER_RGBX,
> >   					      val & PLANE_CTL_ALPHA_MASK);
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		val = I915_READ(PLANE_BUF_CFG(pipe, plane_id));
> > -		skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> > +		skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> >   	} else {
> >   		val = I915_READ(PLANE_BUF_CFG(pipe, plane_id));
> >   		val2 = I915_READ(PLANE_NV12_BUF_CFG(pipe, plane_id));
> > @@ -4009,8 +4009,8 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv,
> >   		if (is_planar_yuv_format(fourcc))
> >   			swap(val, val2);
> > -		skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val);
> > -		skl_ddb_entry_init_from_hw(dev_priv, ddb_uv, val2);
> > +		skl_ddb_entry_init_from_hw(i915, ddb_y, val);
> > +		skl_ddb_entry_init_from_hw(i915, ddb_uv, val2);
> >   	}
> >   }
> > @@ -4018,30 +4018,30 @@ void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc,
> >   			       struct skl_ddb_entry *ddb_y,
> >   			       struct skl_ddb_entry *ddb_uv)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	enum pipe pipe = crtc->pipe;
> >   	intel_wakeref_t wakeref;
> >   	enum plane_id plane_id;
> >   	power_domain = POWER_DOMAIN_PIPE(pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return;
> >   	for_each_plane_id_on_crtc(crtc, plane_id)
> > -		skl_ddb_get_hw_plane_state(dev_priv, pipe,
> > +		skl_ddb_get_hw_plane_state(i915, pipe,
> >   					   plane_id,
> >   					   &ddb_y[plane_id],
> >   					   &ddb_uv[plane_id]);
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   }
> > -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv,
> > +void skl_ddb_get_hw_state(struct drm_i915_private *i915,
> >   			  struct skl_ddb_allocation *ddb /* out */)
> >   {
> > -	ddb->enabled_slices = intel_enabled_dbuf_slices_num(dev_priv);
> > +	ddb->enabled_slices = intel_enabled_dbuf_slices_num(i915);
> >   }
> >   /*
> > @@ -4138,7 +4138,7 @@ skl_pipe_downscale_amount(const struct intel_crtc_state *crtc_state)
> >   int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> >   				  struct intel_crtc_state *cstate)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev);
> >   	struct drm_crtc_state *crtc_state = &cstate->base;
> >   	struct drm_atomic_state *state = crtc_state->state;
> >   	struct drm_plane *plane;
> > @@ -4181,7 +4181,7 @@ int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> >   	crtc_clock = crtc_state->adjusted_mode.crtc_clock;
> >   	dotclk = to_intel_atomic_state(state)->cdclk.logical.cdclk;
> > -	if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10)
> > +	if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10)
> >   		dotclk *= 2;
> >   	pipe_max_pixel_rate = div_round_up_u32_fixed16(dotclk, pipe_downscale);
> > @@ -4341,7 +4341,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   {
> >   	struct drm_atomic_state *state = cstate->base.state;
> >   	struct drm_crtc *crtc = cstate->base.crtc;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct skl_ddb_entry *alloc = &cstate->wm.skl.ddb;
> >   	u16 alloc_size, start = 0;
> > @@ -4367,7 +4367,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   		return 0;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		total_data_rate =
> >   			icl_get_total_relative_data_rate(cstate,
> >   							 plane_data_rate);
> > @@ -4378,7 +4378,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   							 uv_plane_data_rate);
> > -	skl_ddb_get_pipe_allocation_limits(dev_priv, cstate, total_data_rate,
> > +	skl_ddb_get_pipe_allocation_limits(i915, cstate, total_data_rate,
> >   					   ddb, alloc, &num_active);
> >   	alloc_size = skl_ddb_entry_size(alloc);
> >   	if (alloc_size == 0)
> > @@ -4398,7 +4398,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   	 * Find the highest watermark level for which we can satisfy the block
> >   	 * requirement of active planes.
> >   	 */
> > -	for (level = ilk_wm_max_level(dev_priv); level >= 0; level--) {
> > +	for (level = ilk_wm_max_level(i915); level >= 0; level--) {
> >   		blocks = 0;
> >   		for_each_plane_id_on_crtc(intel_crtc, plane_id) {
> >   			const struct skl_plane_wm *wm =
> > @@ -4484,7 +4484,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   			continue;
> >   		/* Gen11+ uses a separate plane for UV watermarks */
> > -		WARN_ON(INTEL_GEN(dev_priv) >= 11 && uv_total[plane_id]);
> > +		WARN_ON(INTEL_GEN(i915) >= 11 && uv_total[plane_id]);
> >   		/* Leave disabled planes at (0,0) */
> >   		if (total[plane_id]) {
> > @@ -4506,7 +4506,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   	 * all levels as "enabled."  Go back now and disable the ones
> >   	 * that aren't actually possible.
> >   	 */
> > -	for (level++; level <= ilk_wm_max_level(dev_priv); level++) {
> > +	for (level++; level <= ilk_wm_max_level(i915); level++) {
> >   		for_each_plane_id_on_crtc(intel_crtc, plane_id) {
> >   			struct skl_plane_wm *wm =
> >   				&cstate->wm.skl.optimal.planes[plane_id];
> > @@ -4531,7 +4531,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >   			 * Wa_1408961008:icl, ehl
> >   			 * Underruns with WM1+ disabled
> >   			 */
> > -			if (IS_GEN(dev_priv, 11) &&
> > +			if (IS_GEN(i915, 11) &&
> >   			    level == 1 && wm->wm[0].plane_en) {
> >   				wm->wm[level].plane_res_b = wm->wm[0].plane_res_b;
> >   				wm->wm[level].plane_res_l = wm->wm[0].plane_res_l;
> > @@ -4562,7 +4562,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate,
> >    * 2xcdclk is 1350 MHz and the pixel rate should never exceed that.
> >   */
> >   static uint_fixed_16_16_t
> > -skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate,
> > +skl_wm_method1(const struct drm_i915_private *i915, u32 pixel_rate,
> >   	       u8 cpp, u32 latency, u32 dbuf_block_size)
> >   {
> >   	u32 wm_intermediate_val;
> > @@ -4574,7 +4574,7 @@ skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate,
> >   	wm_intermediate_val = latency * pixel_rate * cpp;
> >   	ret = div_fixed16(wm_intermediate_val, 1000 * dbuf_block_size);
> > -	if (INTEL_GEN(dev_priv) >= 10)
> > +	if (INTEL_GEN(i915) >= 10)
> >   		ret = add_fixed16_u32(ret, 1);
> >   	return ret;
> > @@ -4648,7 +4648,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> >   		      int color_plane)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	u32 interm_pbpl;
> >   	/* only planar format has two planes */
> > @@ -4673,7 +4673,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> >   	wp->cpp = format->cpp[color_plane];
> >   	wp->plane_pixel_rate = plane_pixel_rate;
> > -	if (INTEL_GEN(dev_priv) >= 11 &&
> > +	if (INTEL_GEN(i915) >= 11 &&
> >   	    modifier == I915_FORMAT_MOD_Yf_TILED  && wp->cpp == 1)
> >   		wp->dbuf_block_size = 256;
> >   	else
> > @@ -4698,7 +4698,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> >   		wp->y_min_scanlines = 4;
> >   	}
> > -	if (skl_needs_memory_bw_wa(dev_priv))
> > +	if (skl_needs_memory_bw_wa(i915))
> >   		wp->y_min_scanlines *= 2;
> >   	wp->plane_bytes_per_line = wp->width * wp->cpp;
> > @@ -4707,12 +4707,12 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state,
> >   					   wp->y_min_scanlines,
> >   					   wp->dbuf_block_size);
> > -		if (INTEL_GEN(dev_priv) >= 10)
> > +		if (INTEL_GEN(i915) >= 10)
> >   			interm_pbpl++;
> >   		wp->plane_blocks_per_line = div_fixed16(interm_pbpl,
> >   							wp->y_min_scanlines);
> > -	} else if (wp->x_tiled && IS_GEN(dev_priv, 9)) {
> > +	} else if (wp->x_tiled && IS_GEN(i915, 9)) {
> >   		interm_pbpl = DIV_ROUND_UP(wp->plane_bytes_per_line,
> >   					   wp->dbuf_block_size);
> >   		wp->plane_blocks_per_line = u32_to_fixed16(interm_pbpl);
> > @@ -4758,9 +4758,9 @@ skl_compute_plane_wm_params(const struct intel_crtc_state *crtc_state,
> >   				     wp, color_plane);
> >   }
> > -static bool skl_wm_has_lines(struct drm_i915_private *dev_priv, int level)
> > +static bool skl_wm_has_lines(struct drm_i915_private *i915, int level)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		return true;
> >   	/* The number of lines are ignored for the level 0 watermark. */
> > @@ -4773,8 +4773,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   				 const struct skl_wm_level *result_prev,
> >   				 struct skl_wm_level *result /* out */)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> > -	u32 latency = dev_priv->wm.skl_latency[level];
> > +	struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> > +	u32 latency = i915->wm.skl_latency[level];
> >   	uint_fixed_16_16_t method1, method2;
> >   	uint_fixed_16_16_t selected_result;
> >   	u32 res_blocks, res_lines, min_ddb_alloc = 0;
> > @@ -4789,14 +4789,14 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   	 * WaIncreaseLatencyIPCEnabled: kbl,cfl
> >   	 * Display WA #1141: kbl,cfl
> >   	 */
> > -	if ((IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) ||
> > -	    dev_priv->ipc_enabled)
> > +	if ((IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) ||
> > +	    i915->ipc_enabled)
> >   		latency += 4;
> > -	if (skl_needs_memory_bw_wa(dev_priv) && wp->x_tiled)
> > +	if (skl_needs_memory_bw_wa(i915) && wp->x_tiled)
> >   		latency += 15;
> > -	method1 = skl_wm_method1(dev_priv, wp->plane_pixel_rate,
> > +	method1 = skl_wm_method1(i915, wp->plane_pixel_rate,
> >   				 wp->cpp, latency, wp->dbuf_block_size);
> >   	method2 = skl_wm_method2(wp->plane_pixel_rate,
> >   				 cstate->base.adjusted_mode.crtc_htotal,
> > @@ -4811,8 +4811,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   		     (wp->plane_bytes_per_line / wp->dbuf_block_size < 1)) {
> >   			selected_result = method2;
> >   		} else if (latency >= wp->linetime_us) {
> > -			if (IS_GEN(dev_priv, 9) &&
> > -			    !IS_GEMINILAKE(dev_priv))
> > +			if (IS_GEN(i915, 9) &&
> > +			    !IS_GEMINILAKE(i915))
> >   				selected_result = min_fixed16(method1, method2);
> >   			else
> >   				selected_result = method2;
> > @@ -4825,7 +4825,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   	res_lines = div_round_up_fixed16(selected_result,
> >   					 wp->plane_blocks_per_line);
> > -	if (IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv)) {
> > +	if (IS_GEN9_BC(i915) || IS_BROXTON(i915)) {
> >   		/* Display WA #1125: skl,bxt,kbl */
> >   		if (level == 0 && wp->rc_surface)
> >   			res_blocks +=
> > @@ -4852,7 +4852,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   		}
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > +	if (INTEL_GEN(i915) >= 11) {
> >   		if (wp->y_tiled) {
> >   			int extra_lines;
> > @@ -4870,7 +4870,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate,
> >   		}
> >   	}
> > -	if (!skl_wm_has_lines(dev_priv, level))
> > +	if (!skl_wm_has_lines(i915, level))
> >   		res_lines = 0;
> >   	if (res_lines > 31) {
> > @@ -4897,8 +4897,8 @@ skl_compute_wm_levels(const struct intel_crtc_state *cstate,
> >   		      const struct skl_wm_params *wm_params,
> >   		      struct skl_wm_level *levels)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	struct skl_wm_level *result_prev = &levels[0];
> >   	for (level = 0; level <= max_level; level++) {
> > @@ -4915,7 +4915,7 @@ static u32
> >   skl_compute_linetime_wm(const struct intel_crtc_state *cstate)
> >   {
> >   	struct drm_atomic_state *state = cstate->base.state;
> > -	struct drm_i915_private *dev_priv = to_i915(state->dev);
> > +	struct drm_i915_private *i915 = to_i915(state->dev);
> >   	uint_fixed_16_16_t linetime_us;
> >   	u32 linetime_wm;
> > @@ -4923,7 +4923,7 @@ skl_compute_linetime_wm(const struct intel_crtc_state *cstate)
> >   	linetime_wm = fixed16_to_u32_round_up(mul_u32_fixed16(8, linetime_us));
> >   	/* Display WA #1135: BXT:ALL GLK:ALL */
> > -	if (IS_GEN9_LP(dev_priv) && dev_priv->ipc_enabled)
> > +	if (IS_GEN9_LP(i915) && i915->ipc_enabled)
> >   		linetime_wm /= 2;
> >   	return linetime_wm;
> > @@ -4934,21 +4934,21 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate,
> >   				      struct skl_plane_wm *wm)
> >   {
> >   	struct drm_device *dev = cstate->base.crtc->dev;
> > -	const struct drm_i915_private *dev_priv = to_i915(dev);
> > +	const struct drm_i915_private *i915 = to_i915(dev);
> >   	u16 trans_min, trans_y_tile_min;
> >   	const u16 trans_amount = 10; /* This is configurable amount */
> >   	u16 wm0_sel_res_b, trans_offset_b, res_blocks;
> >   	/* Transition WM are not recommended by HW team for GEN9 */
> > -	if (INTEL_GEN(dev_priv) <= 9)
> > +	if (INTEL_GEN(i915) <= 9)
> >   		return;
> >   	/* Transition WM don't make any sense if ipc is disabled */
> > -	if (!dev_priv->ipc_enabled)
> > +	if (!i915->ipc_enabled)
> >   		return;
> >   	trans_min = 14;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		trans_min = 4;
> >   	trans_offset_b = trans_min + trans_amount;
> > @@ -4974,7 +4974,7 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate,
> >   		res_blocks = wm0_sel_res_b + trans_offset_b;
> >   		/* WA BUG:1938466 add one block for non y-tile planes */
> > -		if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_A0))
> > +		if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_A0))
> >   			res_blocks += 1;
> >   	}
> > @@ -5093,7 +5093,7 @@ static int icl_build_plane_wm(struct intel_crtc_state *crtc_state,
> >   static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> >   	struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal;
> >   	struct drm_crtc_state *crtc_state = &cstate->base;
> >   	struct drm_plane *plane;
> > @@ -5110,7 +5110,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> >   		const struct intel_plane_state *intel_pstate =
> >   						to_intel_plane_state(pstate);
> > -		if (INTEL_GEN(dev_priv) >= 11)
> > +		if (INTEL_GEN(i915) >= 11)
> >   			ret = icl_build_plane_wm(cstate, intel_pstate);
> >   		else
> >   			ret = skl_build_plane_wm(cstate, intel_pstate);
> > @@ -5123,7 +5123,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate)
> >   	return 0;
> >   }
> > -static void skl_ddb_entry_write(struct drm_i915_private *dev_priv,
> > +static void skl_ddb_entry_write(struct drm_i915_private *i915,
> >   				i915_reg_t reg,
> >   				const struct skl_ddb_entry *entry)
> >   {
> > @@ -5133,7 +5133,7 @@ static void skl_ddb_entry_write(struct drm_i915_private *dev_priv,
> >   		I915_WRITE_FW(reg, 0);
> >   }
> > -static void skl_write_wm_level(struct drm_i915_private *dev_priv,
> > +static void skl_write_wm_level(struct drm_i915_private *i915,
> >   			       i915_reg_t reg,
> >   			       const struct skl_wm_level *level)
> >   {
> > @@ -5152,8 +5152,8 @@ static void skl_write_wm_level(struct drm_i915_private *dev_priv,
> >   void skl_write_plane_wm(struct intel_plane *plane,
> >   			const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	enum plane_id plane_id = plane->id;
> >   	enum pipe pipe = plane->pipe;
> >   	const struct skl_plane_wm *wm =
> > @@ -5164,14 +5164,14 @@ void skl_write_plane_wm(struct intel_plane *plane,
> >   		&crtc_state->wm.skl.plane_ddb_uv[plane_id];
> >   	for (level = 0; level <= max_level; level++) {
> > -		skl_write_wm_level(dev_priv, PLANE_WM(pipe, plane_id, level),
> > +		skl_write_wm_level(i915, PLANE_WM(pipe, plane_id, level),
> >   				   &wm->wm[level]);
> >   	}
> > -	skl_write_wm_level(dev_priv, PLANE_WM_TRANS(pipe, plane_id),
> > +	skl_write_wm_level(i915, PLANE_WM_TRANS(pipe, plane_id),
> >   			   &wm->trans_wm);
> > -	if (INTEL_GEN(dev_priv) >= 11) {
> > -		skl_ddb_entry_write(dev_priv,
> > +	if (INTEL_GEN(i915) >= 11) {
> > +		skl_ddb_entry_write(i915,
> >   				    PLANE_BUF_CFG(pipe, plane_id), ddb_y);
> >   		return;
> >   	}
> > @@ -5179,17 +5179,17 @@ void skl_write_plane_wm(struct intel_plane *plane,
> >   	if (wm->is_planar)
> >   		swap(ddb_y, ddb_uv);
> > -	skl_ddb_entry_write(dev_priv,
> > +	skl_ddb_entry_write(i915,
> >   			    PLANE_BUF_CFG(pipe, plane_id), ddb_y);
> > -	skl_ddb_entry_write(dev_priv,
> > +	skl_ddb_entry_write(i915,
> >   			    PLANE_NV12_BUF_CFG(pipe, plane_id), ddb_uv);
> >   }
> >   void skl_write_cursor_wm(struct intel_plane *plane,
> >   			 const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	enum plane_id plane_id = plane->id;
> >   	enum pipe pipe = plane->pipe;
> >   	const struct skl_plane_wm *wm =
> > @@ -5198,12 +5198,12 @@ void skl_write_cursor_wm(struct intel_plane *plane,
> >   		&crtc_state->wm.skl.plane_ddb_y[plane_id];
> >   	for (level = 0; level <= max_level; level++) {
> > -		skl_write_wm_level(dev_priv, CUR_WM(pipe, level),
> > +		skl_write_wm_level(i915, CUR_WM(pipe, level),
> >   				   &wm->wm[level]);
> >   	}
> > -	skl_write_wm_level(dev_priv, CUR_WM_TRANS(pipe), &wm->trans_wm);
> > +	skl_write_wm_level(i915, CUR_WM_TRANS(pipe), &wm->trans_wm);
> > -	skl_ddb_entry_write(dev_priv, CUR_BUF_CFG(pipe), ddb);
> > +	skl_ddb_entry_write(i915, CUR_BUF_CFG(pipe), ddb);
> >   }
> >   bool skl_wm_level_equals(const struct skl_wm_level *l1,
> > @@ -5215,11 +5215,11 @@ bool skl_wm_level_equals(const struct skl_wm_level *l1,
> >   		l1->plane_res_b == l2->plane_res_b;
> >   }
> > -static bool skl_plane_wm_equals(struct drm_i915_private *dev_priv,
> > +static bool skl_plane_wm_equals(struct drm_i915_private *i915,
> >   				const struct skl_plane_wm *wm1,
> >   				const struct skl_plane_wm *wm2)
> >   {
> > -	int level, max_level = ilk_wm_max_level(dev_priv);
> > +	int level, max_level = ilk_wm_max_level(i915);
> >   	for (level = 0; level <= max_level; level++) {
> >   		if (!skl_wm_level_equals(&wm1->wm[level], &wm2->wm[level]) ||
> > @@ -5234,11 +5234,11 @@ static bool skl_pipe_wm_equals(struct intel_crtc *crtc,
> >   			       const struct skl_pipe_wm *wm1,
> >   			       const struct skl_pipe_wm *wm2)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum plane_id plane_id;
> >   	for_each_plane_id_on_crtc(crtc, plane_id) {
> > -		if (!skl_plane_wm_equals(dev_priv,
> > +		if (!skl_plane_wm_equals(i915,
> >   					 &wm1->planes[plane_id],
> >   					 &wm2->planes[plane_id]))
> >   			return false;
> > @@ -5287,10 +5287,10 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state,
> >   {
> >   	struct intel_atomic_state *state = to_intel_atomic_state(new_crtc_state->base.state);
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	struct intel_plane *plane;
> > -	for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +	for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   		struct intel_plane_state *plane_state;
> >   		enum plane_id plane_id = plane->id;
> > @@ -5313,14 +5313,14 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state,
> >   static int
> >   skl_compute_ddb(struct intel_atomic_state *state)
> >   {
> > -	const struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	const struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct skl_ddb_allocation *ddb = &state->wm_results.ddb;
> >   	struct intel_crtc_state *old_crtc_state;
> >   	struct intel_crtc_state *new_crtc_state;
> >   	struct intel_crtc *crtc;
> >   	int ret, i;
> > -	memcpy(ddb, &dev_priv->wm.skl_hw.ddb, sizeof(*ddb));
> > +	memcpy(ddb, &i915->wm.skl_hw.ddb, sizeof(*ddb));
> >   	for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state,
> >   					    new_crtc_state, i) {
> > @@ -5345,7 +5345,7 @@ static char enast(bool enable)
> >   static void
> >   skl_print_wm_changes(struct intel_atomic_state *state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	const struct intel_crtc_state *old_crtc_state;
> >   	const struct intel_crtc_state *new_crtc_state;
> >   	struct intel_plane *plane;
> > @@ -5362,7 +5362,7 @@ skl_print_wm_changes(struct intel_atomic_state *state)
> >   		old_pipe_wm = &old_crtc_state->wm.skl.optimal;
> >   		new_pipe_wm = &new_crtc_state->wm.skl.optimal;
> > -		for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +		for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   			enum plane_id plane_id = plane->id;
> >   			const struct skl_ddb_entry *old, *new;
> > @@ -5378,14 +5378,14 @@ skl_print_wm_changes(struct intel_atomic_state *state)
> >   				      skl_ddb_entry_size(old), skl_ddb_entry_size(new));
> >   		}
> > -		for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +		for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   			enum plane_id plane_id = plane->id;
> >   			const struct skl_plane_wm *old_wm, *new_wm;
> >   			old_wm = &old_pipe_wm->planes[plane_id];
> >   			new_wm = &new_pipe_wm->planes[plane_id];
> > -			if (skl_plane_wm_equals(dev_priv, old_wm, new_wm))
> > +			if (skl_plane_wm_equals(i915, old_wm, new_wm))
> >   				continue;
> >   			DRM_DEBUG_KMS("[PLANE:%d:%s]   level %cwm0,%cwm1,%cwm2,%cwm3,%cwm4,%cwm5,%cwm6,%cwm7,%ctwm"
> > @@ -5460,7 +5460,7 @@ static int
> >   skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> >   {
> >   	struct drm_device *dev = state->base.dev;
> > -	const struct drm_i915_private *dev_priv = to_i915(dev);
> > +	const struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *crtc;
> >   	struct intel_crtc_state *crtc_state;
> >   	u32 realloc_pipes = pipes_modified(state);
> > @@ -5470,14 +5470,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> >   	 * When we distrust bios wm we always need to recompute to set the
> >   	 * expected DDB allocations for each CRTC.
> >   	 */
> > -	if (dev_priv->wm.distrust_bios_wm)
> > +	if (i915->wm.distrust_bios_wm)
> >   		(*changed) = true;
> >   	/*
> >   	 * If this transaction isn't actually touching any CRTC's, don't
> >   	 * bother with watermark calculation.  Note that if we pass this
> >   	 * test, we're guaranteed to hold at least one CRTC state mutex,
> > -	 * which means we can safely use values like dev_priv->active_crtcs
> > +	 * which means we can safely use values like i915->active_crtcs
> >   	 * since any racing commits that want to update them would need to
> >   	 * hold _all_ CRTC state mutexes.
> >   	 */
> > @@ -5493,7 +5493,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> >   	 * pretend that all pipes switched active status so that we'll
> >   	 * ensure a full DDB recompute.
> >   	 */
> > -	if (dev_priv->wm.distrust_bios_wm) {
> > +	if (i915->wm.distrust_bios_wm) {
> >   		ret = drm_modeset_lock(&dev->mode_config.connection_mutex,
> >   				       state->base.acquire_ctx);
> >   		if (ret)
> > @@ -5508,7 +5508,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> >   		 * on the first commit too.
> >   		 */
> >   		if (!state->modeset)
> > -			state->active_crtcs = dev_priv->active_crtcs;
> > +			state->active_crtcs = i915->active_crtcs;
> >   	}
> >   	/*
> > @@ -5567,14 +5567,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed)
> >   static int skl_wm_add_affected_planes(struct intel_atomic_state *state,
> >   				      struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct intel_crtc_state *old_crtc_state =
> >   		intel_atomic_get_old_crtc_state(state, crtc);
> >   	struct intel_crtc_state *new_crtc_state =
> >   		intel_atomic_get_new_crtc_state(state, crtc);
> >   	struct intel_plane *plane;
> > -	for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) {
> > +	for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) {
> >   		struct intel_plane_state *plane_state;
> >   		enum plane_id plane_id = plane->id;
> > @@ -5587,7 +5587,7 @@ static int skl_wm_add_affected_planes(struct intel_atomic_state *state,
> >   		 * with the software state.
> >   		 */
> >   		if (!drm_atomic_crtc_needs_modeset(&new_crtc_state->base) &&
> > -		    skl_plane_wm_equals(dev_priv,
> > +		    skl_plane_wm_equals(i915,
> >   					&old_crtc_state->wm.skl.optimal.planes[plane_id],
> >   					&new_crtc_state->wm.skl.optimal.planes[plane_id]))
> >   			continue;
> > @@ -5654,7 +5654,7 @@ static void skl_atomic_update_crtc_wm(struct intel_atomic_state *state,
> >   				      struct intel_crtc_state *cstate)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(cstate->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(state->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(state->base.dev);
> >   	struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal;
> >   	enum pipe pipe = crtc->pipe;
> > @@ -5669,27 +5669,27 @@ static void skl_initial_wm(struct intel_atomic_state *state,
> >   {
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> >   	struct drm_device *dev = intel_crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct skl_ddb_values *results = &state->wm_results;
> >   	if ((results->dirty_pipes & drm_crtc_mask(&intel_crtc->base)) == 0)
> >   		return;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	if (cstate->base.active_changed)
> >   		skl_atomic_update_crtc_wm(state, cstate);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> > -static void ilk_compute_wm_config(struct drm_i915_private *dev_priv,
> > +static void ilk_compute_wm_config(struct drm_i915_private *i915,
> >   				  struct intel_wm_config *config)
> >   {
> >   	struct intel_crtc *crtc;
> >   	/* Compute the currently _active_ config */
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		const struct intel_pipe_wm *wm = &crtc->wm.active.ilk;
> >   		if (!wm->pipe_enabled)
> > @@ -5701,7 +5701,7 @@ static void ilk_compute_wm_config(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> > +static void ilk_program_watermarks(struct drm_i915_private *i915)
> >   {
> >   	struct intel_pipe_wm lp_wm_1_2 = {}, lp_wm_5_6 = {}, *best_lp_wm;
> >   	struct ilk_wm_maximums max;
> > @@ -5709,18 +5709,18 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> >   	struct ilk_wm_values results = {};
> >   	enum intel_ddb_partitioning partitioning;
> > -	ilk_compute_wm_config(dev_priv, &config);
> > +	ilk_compute_wm_config(i915, &config);
> > -	ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_1_2, &max);
> > -	ilk_wm_merge(dev_priv, &config, &max, &lp_wm_1_2);
> > +	ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_1_2, &max);
> > +	ilk_wm_merge(i915, &config, &max, &lp_wm_1_2);
> >   	/* 5/6 split only in single pipe config on IVB+ */
> > -	if (INTEL_GEN(dev_priv) >= 7 &&
> > +	if (INTEL_GEN(i915) >= 7 &&
> >   	    config.num_pipes_active == 1 && config.sprites_enabled) {
> > -		ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_5_6, &max);
> > -		ilk_wm_merge(dev_priv, &config, &max, &lp_wm_5_6);
> > +		ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_5_6, &max);
> > +		ilk_wm_merge(i915, &config, &max, &lp_wm_5_6);
> > -		best_lp_wm = ilk_find_best_result(dev_priv, &lp_wm_1_2, &lp_wm_5_6);
> > +		best_lp_wm = ilk_find_best_result(i915, &lp_wm_1_2, &lp_wm_5_6);
> >   	} else {
> >   		best_lp_wm = &lp_wm_1_2;
> >   	}
> > @@ -5728,35 +5728,35 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv)
> >   	partitioning = (best_lp_wm == &lp_wm_1_2) ?
> >   		       INTEL_DDB_PART_1_2 : INTEL_DDB_PART_5_6;
> > -	ilk_compute_wm_results(dev_priv, best_lp_wm, partitioning, &results);
> > +	ilk_compute_wm_results(i915, best_lp_wm, partitioning, &results);
> > -	ilk_write_wm_values(dev_priv, &results);
> > +	ilk_write_wm_values(i915, &results);
> >   }
> >   static void ilk_initial_watermarks(struct intel_atomic_state *state,
> >   				   struct intel_crtc_state *cstate)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	intel_crtc->wm.active.ilk = cstate->wm.ilk.intermediate;
> > -	ilk_program_watermarks(dev_priv);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	ilk_program_watermarks(i915);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   static void ilk_optimize_watermarks(struct intel_atomic_state *state,
> >   				    struct intel_crtc_state *cstate)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc);
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> >   	if (cstate->wm.need_postvbl_update) {
> >   		intel_crtc->wm.active.ilk = cstate->wm.ilk.optimal;
> > -		ilk_program_watermarks(dev_priv);
> > +		ilk_program_watermarks(i915);
> >   	}
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   static inline void skl_wm_level_from_reg_val(u32 val,
> > @@ -5772,13 +5772,13 @@ static inline void skl_wm_level_from_reg_val(u32 val,
> >   void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> >   			      struct skl_pipe_wm *out)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	int level, max_level;
> >   	enum plane_id plane_id;
> >   	u32 val;
> > -	max_level = ilk_wm_max_level(dev_priv);
> > +	max_level = ilk_wm_max_level(i915);
> >   	for_each_plane_id_on_crtc(crtc, plane_id) {
> >   		struct skl_plane_wm *wm = &out->planes[plane_id];
> > @@ -5806,15 +5806,15 @@ void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> >   	out->linetime = I915_READ(PIPE_WM_LINETIME(pipe));
> >   }
> > -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv)
> > +void skl_wm_get_hw_state(struct drm_i915_private *i915)
> >   {
> > -	struct skl_ddb_values *hw = &dev_priv->wm.skl_hw;
> > -	struct skl_ddb_allocation *ddb = &dev_priv->wm.skl_hw.ddb;
> > +	struct skl_ddb_values *hw = &i915->wm.skl_hw;
> > +	struct skl_ddb_allocation *ddb = &i915->wm.skl_hw.ddb;
> >   	struct intel_crtc *crtc;
> >   	struct intel_crtc_state *cstate;
> > -	skl_ddb_get_hw_state(dev_priv, ddb);
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	skl_ddb_get_hw_state(i915, ddb);
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		cstate = to_intel_crtc_state(crtc->base.state);
> >   		skl_pipe_wm_get_hw_state(crtc, &cstate->wm.skl.optimal);
> > @@ -5823,17 +5823,17 @@ void skl_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >   			hw->dirty_pipes |= drm_crtc_mask(&crtc->base);
> >   	}
> > -	if (dev_priv->active_crtcs) {
> > +	if (i915->active_crtcs) {
> >   		/* Fully recompute DDB on first atomic commit */
> > -		dev_priv->wm.distrust_bios_wm = true;
> > +		i915->wm.distrust_bios_wm = true;
> >   	}
> >   }
> >   static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> >   {
> >   	struct drm_device *dev = crtc->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct ilk_wm_values *hw = &dev_priv->wm.hw;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct ilk_wm_values *hw = &i915->wm.hw;
> >   	struct intel_crtc_state *cstate = to_intel_crtc_state(crtc->base.state);
> >   	struct intel_pipe_wm *active = &cstate->wm.ilk.optimal;
> >   	enum pipe pipe = crtc->pipe;
> > @@ -5844,7 +5844,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> >   	};
> >   	hw->wm_pipe[pipe] = I915_READ(wm0_pipe_reg[pipe]);
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		hw->wm_linetime[pipe] = I915_READ(PIPE_WM_LINETIME(pipe));
> >   	memset(active, 0, sizeof(*active));
> > @@ -5866,7 +5866,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> >   		active->wm[0].cur_val = tmp & WM0_PIPE_CURSOR_MASK;
> >   		active->linetime = hw->wm_linetime[pipe];
> >   	} else {
> > -		int level, max_level = ilk_wm_max_level(dev_priv);
> > +		int level, max_level = ilk_wm_max_level(i915);
> >   		/*
> >   		 * For inactive pipes, all watermark levels
> > @@ -5885,7 +5885,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc)
> >   #define _FW_WM_VLV(value, plane) \
> >   	(((value) & DSPFW_ ## plane ## _MASK_VLV) >> DSPFW_ ## plane ## _SHIFT)
> > -static void g4x_read_wm_values(struct drm_i915_private *dev_priv,
> > +static void g4x_read_wm_values(struct drm_i915_private *i915,
> >   			       struct g4x_wm_values *wm)
> >   {
> >   	u32 tmp;
> > @@ -5911,13 +5911,13 @@ static void g4x_read_wm_values(struct drm_i915_private *dev_priv,
> >   	wm->hpll.plane = _FW_WM(tmp, HPLL_SR);
> >   }
> > -static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> > +static void vlv_read_wm_values(struct drm_i915_private *i915,
> >   			       struct vlv_wm_values *wm)
> >   {
> >   	enum pipe pipe;
> >   	u32 tmp;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		tmp = I915_READ(VLV_DDL(pipe));
> >   		wm->ddl[pipe].plane[PLANE_PRIMARY] =
> > @@ -5944,7 +5944,7 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> >   	tmp = I915_READ(DSPFW3);
> >   	wm->sr.cursor = _FW_WM(tmp, CURSOR_SR);
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		tmp = I915_READ(DSPFW7_CHV);
> >   		wm->pipe[PIPE_B].plane[PLANE_SPRITE1] = _FW_WM_VLV(tmp, SPRITED);
> >   		wm->pipe[PIPE_B].plane[PLANE_SPRITE0] = _FW_WM_VLV(tmp, SPRITEC);
> > @@ -5987,16 +5987,16 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv,
> >   #undef _FW_WM
> >   #undef _FW_WM_VLV
> > -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv)
> > +void g4x_wm_get_hw_state(struct drm_i915_private *i915)
> >   {
> > -	struct g4x_wm_values *wm = &dev_priv->wm.g4x;
> > +	struct g4x_wm_values *wm = &i915->wm.g4x;
> >   	struct intel_crtc *crtc;
> > -	g4x_read_wm_values(dev_priv, wm);
> > +	g4x_read_wm_values(i915, wm);
> >   	wm->cxsr = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN;
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		struct g4x_wm_state *active = &crtc->wm.active.g4x;
> > @@ -6071,16 +6071,16 @@ void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >   		      yesno(wm->cxsr), yesno(wm->hpll_en), yesno(wm->fbc_en));
> >   }
> > -void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> > +void g4x_wm_sanitize(struct drm_i915_private *i915)
> >   {
> >   	struct intel_plane *plane;
> >   	struct intel_crtc *crtc;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> > -	for_each_intel_plane(&dev_priv->drm, plane) {
> > +	for_each_intel_plane(&i915->drm, plane) {
> >   		struct intel_crtc *crtc =
> > -			intel_get_crtc_for_pipe(dev_priv, plane->pipe);
> > +			intel_get_crtc_for_pipe(i915, plane->pipe);
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		struct intel_plane_state *plane_state =
> > @@ -6113,7 +6113,7 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> >   		}
> >   	}
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> > @@ -6122,26 +6122,26 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv)
> >   		crtc->wm.active.g4x = crtc_state->wm.g4x.optimal;
> >   	}
> > -	g4x_program_watermarks(dev_priv);
> > +	g4x_program_watermarks(i915);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> > -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> > +void vlv_wm_get_hw_state(struct drm_i915_private *i915)
> >   {
> > -	struct vlv_wm_values *wm = &dev_priv->wm.vlv;
> > +	struct vlv_wm_values *wm = &i915->wm.vlv;
> >   	struct intel_crtc *crtc;
> >   	u32 val;
> > -	vlv_read_wm_values(dev_priv, wm);
> > +	vlv_read_wm_values(i915, wm);
> >   	wm->cxsr = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN;
> >   	wm->level = VLV_WM_LEVEL_PM2;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > -		vlv_punit_get(dev_priv);
> > +	if (IS_CHERRYVIEW(i915)) {
> > +		vlv_punit_get(i915);
> > -		val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM);
> > +		val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM);
> >   		if (val & DSP_MAXFIFO_PM5_ENABLE)
> >   			wm->level = VLV_WM_LEVEL_PM5;
> > @@ -6154,25 +6154,25 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >   		 * HIGH/LOW bits so that we don't actually change
> >   		 * the current state.
> >   		 */
> > -		val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> > +		val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> >   		val |= FORCE_DDR_FREQ_REQ_ACK;
> > -		vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val);
> > +		vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val);
> > -		if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) &
> > +		if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) &
> >   			      FORCE_DDR_FREQ_REQ_ACK) == 0, 3)) {
> >   			DRM_DEBUG_KMS("Punit not acking DDR DVFS request, "
> >   				      "assuming DDR DVFS is disabled\n");
> > -			dev_priv->wm.max_level = VLV_WM_LEVEL_PM5;
> > +			i915->wm.max_level = VLV_WM_LEVEL_PM5;
> >   		} else {
> > -			val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
> > +			val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2);
> >   			if ((val & FORCE_DDR_HIGH_FREQ) == 0)
> >   				wm->level = VLV_WM_LEVEL_DDR_DVFS;
> >   		}
> > -		vlv_punit_put(dev_priv);
> > +		vlv_punit_put(i915);
> >   	}
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		struct vlv_wm_state *active = &crtc->wm.active.vlv;
> > @@ -6224,16 +6224,16 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >   		      wm->sr.plane, wm->sr.cursor, wm->level, wm->cxsr);
> >   }
> > -void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> > +void vlv_wm_sanitize(struct drm_i915_private *i915)
> >   {
> >   	struct intel_plane *plane;
> >   	struct intel_crtc *crtc;
> > -	mutex_lock(&dev_priv->wm.wm_mutex);
> > +	mutex_lock(&i915->wm.wm_mutex);
> > -	for_each_intel_plane(&dev_priv->drm, plane) {
> > +	for_each_intel_plane(&i915->drm, plane) {
> >   		struct intel_crtc *crtc =
> > -			intel_get_crtc_for_pipe(dev_priv, plane->pipe);
> > +			intel_get_crtc_for_pipe(i915, plane->pipe);
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> >   		struct intel_plane_state *plane_state =
> > @@ -6259,7 +6259,7 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> >   		}
> >   	}
> > -	for_each_intel_crtc(&dev_priv->drm, crtc) {
> > +	for_each_intel_crtc(&i915->drm, crtc) {
> >   		struct intel_crtc_state *crtc_state =
> >   			to_intel_crtc_state(crtc->base.state);
> > @@ -6268,16 +6268,16 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv)
> >   		crtc->wm.active.vlv = crtc_state->wm.vlv.optimal;
> >   	}
> > -	vlv_program_watermarks(dev_priv);
> > +	vlv_program_watermarks(i915);
> > -	mutex_unlock(&dev_priv->wm.wm_mutex);
> > +	mutex_unlock(&i915->wm.wm_mutex);
> >   }
> >   /*
> >    * FIXME should probably kill this and improve
> >    * the real watermark readout/sanitation instead
> >    */
> > -static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv)
> > +static void ilk_init_lp_watermarks(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(WM3_LP_ILK, I915_READ(WM3_LP_ILK) & ~WM1_LP_SR_EN);
> >   	I915_WRITE(WM2_LP_ILK, I915_READ(WM2_LP_ILK) & ~WM1_LP_SR_EN);
> > @@ -6289,14 +6289,14 @@ static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv)
> >   	 */
> >   }
> > -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> > +void ilk_wm_get_hw_state(struct drm_i915_private *i915)
> >   {
> > -	struct ilk_wm_values *hw = &dev_priv->wm.hw;
> > +	struct ilk_wm_values *hw = &i915->wm.hw;
> >   	struct intel_crtc *crtc;
> > -	ilk_init_lp_watermarks(dev_priv);
> > +	ilk_init_lp_watermarks(i915);
> > -	for_each_intel_crtc(&dev_priv->drm, crtc)
> > +	for_each_intel_crtc(&i915->drm, crtc)
> >   		ilk_pipe_wm_get_hw_state(crtc);
> >   	hw->wm_lp[0] = I915_READ(WM1_LP_ILK);
> > @@ -6304,15 +6304,15 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >   	hw->wm_lp[2] = I915_READ(WM3_LP_ILK);
> >   	hw->wm_lp_spr[0] = I915_READ(WM1S_LP_ILK);
> > -	if (INTEL_GEN(dev_priv) >= 7) {
> > +	if (INTEL_GEN(i915) >= 7) {
> >   		hw->wm_lp_spr[1] = I915_READ(WM2S_LP_IVB);
> >   		hw->wm_lp_spr[2] = I915_READ(WM3S_LP_IVB);
> >   	}
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		hw->partitioning = (I915_READ(WM_MISC) & WM_MISC_DATA_PARTITION_5_6) ?
> >   			INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2;
> > -	else if (IS_IVYBRIDGE(dev_priv))
> > +	else if (IS_IVYBRIDGE(i915))
> >   		hw->partitioning = (I915_READ(DISP_ARB_CTL2) & DISP_DATA_PARTITION_5_6) ?
> >   			INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2;
> > @@ -6355,22 +6355,22 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
> >    */
> >   void intel_update_watermarks(struct intel_crtc *crtc)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	if (dev_priv->display.update_wm)
> > -		dev_priv->display.update_wm(crtc);
> > +	if (i915->display.update_wm)
> > +		i915->display.update_wm(crtc);
> >   }
> > -void intel_enable_ipc(struct drm_i915_private *dev_priv)
> > +void intel_enable_ipc(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	if (!HAS_IPC(dev_priv))
> > +	if (!HAS_IPC(i915))
> >   		return;
> >   	val = I915_READ(DISP_ARB_CTL2);
> > -	if (dev_priv->ipc_enabled)
> > +	if (i915->ipc_enabled)
> >   		val |= DISP_IPC_ENABLE;
> >   	else
> >   		val &= ~DISP_IPC_ENABLE;
> > @@ -6378,27 +6378,27 @@ void intel_enable_ipc(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(DISP_ARB_CTL2, val);
> >   }
> > -static bool intel_can_enable_ipc(struct drm_i915_private *dev_priv)
> > +static bool intel_can_enable_ipc(struct drm_i915_private *i915)
> >   {
> >   	/* Display WA #0477 WaDisableIPC: skl */
> > -	if (IS_SKYLAKE(dev_priv))
> > +	if (IS_SKYLAKE(i915))
> >   		return false;
> >   	/* Display WA #1141: SKL:all KBL:all CFL */
> > -	if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv))
> > -		return dev_priv->dram_info.symmetric_memory;
> > +	if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915))
> > +		return i915->dram_info.symmetric_memory;
> >   	return true;
> >   }
> > -void intel_init_ipc(struct drm_i915_private *dev_priv)
> > +void intel_init_ipc(struct drm_i915_private *i915)
> >   {
> > -	if (!HAS_IPC(dev_priv))
> > +	if (!HAS_IPC(i915))
> >   		return;
> > -	dev_priv->ipc_enabled = intel_can_enable_ipc(dev_priv);
> > +	i915->ipc_enabled = intel_can_enable_ipc(i915);
> > -	intel_enable_ipc(dev_priv);
> > +	intel_enable_ipc(i915);
> >   }
> >   /*
> > @@ -6430,9 +6430,9 @@ bool ironlake_set_drps(struct drm_i915_private *i915, u8 val)
> >   	return true;
> >   }
> > -static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> > +static void ironlake_enable_drps(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u32 rgvmodectl;
> >   	u8 fmax, fmin, fstart, vstart;
> > @@ -6463,12 +6463,12 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> >   	vstart = (intel_uncore_read(uncore, PXVFREQ(fstart)) &
> >   		  PXVFREQ_PX_MASK) >> PXVFREQ_PX_SHIFT;
> > -	dev_priv->ips.fmax = fmax; /* IPS callback will increase this */
> > -	dev_priv->ips.fstart = fstart;
> > +	i915->ips.fmax = fmax; /* IPS callback will increase this */
> > +	i915->ips.fstart = fstart;
> > -	dev_priv->ips.max_delay = fstart;
> > -	dev_priv->ips.min_delay = fmin;
> > -	dev_priv->ips.cur_delay = fstart;
> > +	i915->ips.max_delay = fstart;
> > +	i915->ips.min_delay = fmin;
> > +	i915->ips.cur_delay = fstart;
> >   	DRM_DEBUG_DRIVER("fmax: %d, fmin: %d, fstart: %d\n",
> >   			 fmax, fmin, fstart);
> > @@ -6492,15 +6492,15 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv)
> >   		DRM_ERROR("stuck trying to change perf mode\n");
> >   	mdelay(1);
> > -	ironlake_set_drps(dev_priv, fstart);
> > +	ironlake_set_drps(i915, fstart);
> > -	dev_priv->ips.last_count1 =
> > +	i915->ips.last_count1 =
> >   		intel_uncore_read(uncore, DMIEC) +
> >   		intel_uncore_read(uncore, DDREC) +
> >   		intel_uncore_read(uncore, CSIEC);
> > -	dev_priv->ips.last_time1 = jiffies_to_msecs(jiffies);
> > -	dev_priv->ips.last_count2 = intel_uncore_read(uncore, GFXEC);
> > -	dev_priv->ips.last_time2 = ktime_get_raw_ns();
> > +	i915->ips.last_time1 = jiffies_to_msecs(jiffies);
> > +	i915->ips.last_count2 = intel_uncore_read(uncore, GFXEC);
> > +	i915->ips.last_time2 = ktime_get_raw_ns();
> >   	spin_unlock_irq(&mchdev_lock);
> >   }
> > @@ -6543,9 +6543,9 @@ static void ironlake_disable_drps(struct drm_i915_private *i915)
> >    * ourselves, instead of doing a rmw cycle (which might result in us clearing
> >    * all limits and the gpu stuck at whatever frequency it is at atm).
> >    */
> > -static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> > +static u32 intel_rps_limits(struct drm_i915_private *i915, u8 val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 limits;
> >   	/* Only set the down limit when we've reached the lowest level to avoid
> > @@ -6554,7 +6554,7 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> >   	 * the hw runs at the minimal clock before selecting the desired
> >   	 * frequency, if the down threshold expires in that window we will not
> >   	 * receive a down interrupt. */
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > +	if (INTEL_GEN(i915) >= 9) {
> >   		limits = (rps->max_freq_softlimit) << 23;
> >   		if (val <= rps->min_freq_softlimit)
> >   			limits |= (rps->min_freq_softlimit) << 14;
> > @@ -6567,9 +6567,9 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val)
> >   	return limits;
> >   }
> > -static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> > +static void rps_set_power(struct drm_i915_private *i915, int new_power)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 threshold_up = 0, threshold_down = 0; /* in % */
> >   	u32 ei_up = 0, ei_down = 0;
> > @@ -6614,23 +6614,23 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> >   	/* When byt can survive without system hang with dynamic
> >   	 * sw freq adjustments, this restriction can be lifted.
> >   	 */
> > -	if (IS_VALLEYVIEW(dev_priv))
> > +	if (IS_VALLEYVIEW(i915))
> >   		goto skip_hw_write;
> >   	I915_WRITE(GEN6_RP_UP_EI,
> > -		   GT_INTERVAL_FROM_US(dev_priv, ei_up));
> > +		   GT_INTERVAL_FROM_US(i915, ei_up));
> >   	I915_WRITE(GEN6_RP_UP_THRESHOLD,
> > -		   GT_INTERVAL_FROM_US(dev_priv,
> > +		   GT_INTERVAL_FROM_US(i915,
> >   				       ei_up * threshold_up / 100));
> >   	I915_WRITE(GEN6_RP_DOWN_EI,
> > -		   GT_INTERVAL_FROM_US(dev_priv, ei_down));
> > +		   GT_INTERVAL_FROM_US(i915, ei_down));
> >   	I915_WRITE(GEN6_RP_DOWN_THRESHOLD,
> > -		   GT_INTERVAL_FROM_US(dev_priv,
> > +		   GT_INTERVAL_FROM_US(i915,
> >   				       ei_down * threshold_down / 100));
> >   	I915_WRITE(GEN6_RP_CONTROL,
> > -		   (INTEL_GEN(dev_priv) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) |
> > +		   (INTEL_GEN(i915) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) |
> >   		   GEN6_RP_MEDIA_HW_NORMAL_MODE |
> >   		   GEN6_RP_MEDIA_IS_GFX |
> >   		   GEN6_RP_ENABLE |
> > @@ -6643,9 +6643,9 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power)
> >   	rps->power.down_threshold = threshold_down;
> >   }
> > -static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val)
> > +static void gen6_set_rps_thresholds(struct drm_i915_private *i915, u8 val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	int new_power;
> >   	new_power = rps->power.mode;
> > @@ -6680,7 +6680,7 @@ static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val)
> >   	mutex_lock(&rps->power.mutex);
> >   	if (rps->power.interactive)
> >   		new_power = HIGH_POWER;
> > -	rps_set_power(dev_priv, new_power);
> > +	rps_set_power(i915, new_power);
> >   	mutex_unlock(&rps->power.mutex);
> >   }
> > @@ -6702,9 +6702,9 @@ void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive)
> >   	mutex_unlock(&rps->power.mutex);
> >   }
> > -static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val)
> > +static u32 gen6_rps_pm_mask(struct drm_i915_private *i915, u8 val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 mask = 0;
> >   	/* We use UP_EI_EXPIRED interupts for both up/down in manual mode */
> > @@ -6713,28 +6713,28 @@ static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val)
> >   	if (val < rps->max_freq_softlimit)
> >   		mask |= GEN6_PM_RP_UP_EI_EXPIRED | GEN6_PM_RP_UP_THRESHOLD;
> > -	mask &= dev_priv->pm_rps_events;
> > +	mask &= i915->pm_rps_events;
> > -	return gen6_sanitize_rps_pm_mask(dev_priv, ~mask);
> > +	return gen6_sanitize_rps_pm_mask(i915, ~mask);
> >   }
> >   /* gen6_set_rps is called to update the frequency request, but should also be
> >    * called when the range (min_delay and max_delay) is modified so that we can
> >    * update the GEN6_RP_INTERRUPT_LIMITS register accordingly. */
> > -static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val)
> > +static int gen6_set_rps(struct drm_i915_private *i915, u8 val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/* min/max delay may still have been modified so be sure to
> >   	 * write the limits value.
> >   	 */
> >   	if (val != rps->cur_freq) {
> > -		gen6_set_rps_thresholds(dev_priv, val);
> > +		gen6_set_rps_thresholds(i915, val);
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +		if (INTEL_GEN(i915) >= 9)
> >   			I915_WRITE(GEN6_RPNSWREQ,
> >   				   GEN9_FREQUENCY(val));
> > -		else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +		else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   			I915_WRITE(GEN6_RPNSWREQ,
> >   				   HSW_FREQUENCY(val));
> >   		else
> > @@ -6747,37 +6747,37 @@ static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val)
> >   	/* Make sure we continue to get interrupts
> >   	 * until we hit the minimum or maximum frequencies.
> >   	 */
> > -	I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(dev_priv, val));
> > -	I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val));
> > +	I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(i915, val));
> > +	I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val));
> >   	rps->cur_freq = val;
> > -	trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val));
> > +	trace_intel_gpu_freq_change(intel_gpu_freq(i915, val));
> >   	return 0;
> >   }
> > -static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val)
> > +static int valleyview_set_rps(struct drm_i915_private *i915, u8 val)
> >   {
> >   	int err;
> > -	if (WARN_ONCE(IS_CHERRYVIEW(dev_priv) && (val & 1),
> > +	if (WARN_ONCE(IS_CHERRYVIEW(i915) && (val & 1),
> >   		      "Odd GPU freq value\n"))
> >   		val &= ~1;
> > -	I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val));
> > +	I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val));
> > -	if (val != dev_priv->gt_pm.rps.cur_freq) {
> > -		vlv_punit_get(dev_priv);
> > -		err = vlv_punit_write(dev_priv, PUNIT_REG_GPU_FREQ_REQ, val);
> > -		vlv_punit_put(dev_priv);
> > +	if (val != i915->gt_pm.rps.cur_freq) {
> > +		vlv_punit_get(i915);
> > +		err = vlv_punit_write(i915, PUNIT_REG_GPU_FREQ_REQ, val);
> > +		vlv_punit_put(i915);
> >   		if (err)
> >   			return err;
> > -		gen6_set_rps_thresholds(dev_priv, val);
> > +		gen6_set_rps_thresholds(i915, val);
> >   	}
> > -	dev_priv->gt_pm.rps.cur_freq = val;
> > -	trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val));
> > +	i915->gt_pm.rps.cur_freq = val;
> > +	trace_intel_gpu_freq_change(intel_gpu_freq(i915, val));
> >   	return 0;
> >   }
> > @@ -6789,9 +6789,9 @@ static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val)
> >    * 2. Request idle freq.
> >    * 3. Release Forcewake of Media well.
> >   */
> > -static void vlv_set_rps_idle(struct drm_i915_private *dev_priv)
> > +static void vlv_set_rps_idle(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 val = rps->idle_freq;
> >   	int err;
> > @@ -6810,28 +6810,28 @@ static void vlv_set_rps_idle(struct drm_i915_private *dev_priv)
> >   	 * punit into committing the voltage change) as that takes a lot less
> >   	 * power than the render powerwell.
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_MEDIA);
> > -	err = valleyview_set_rps(dev_priv, val);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_MEDIA);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_MEDIA);
> > +	err = valleyview_set_rps(i915, val);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_MEDIA);
> >   	if (err)
> >   		DRM_ERROR("Failed to set RPS for idle\n");
> >   }
> > -void gen6_rps_busy(struct drm_i915_private *dev_priv)
> > +void gen6_rps_busy(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	mutex_lock(&rps->lock);
> >   	if (rps->enabled) {
> >   		u8 freq;
> > -		if (dev_priv->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED)
> > -			gen6_rps_reset_ei(dev_priv);
> > +		if (i915->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED)
> > +			gen6_rps_reset_ei(i915);
> >   		I915_WRITE(GEN6_PMINTRMSK,
> > -			   gen6_rps_pm_mask(dev_priv, rps->cur_freq));
> > +			   gen6_rps_pm_mask(i915, rps->cur_freq));
> > -		gen6_enable_rps_interrupts(dev_priv);
> > +		gen6_enable_rps_interrupts(i915);
> >   		/* Use the user's desired frequency as a guide, but for better
> >   		 * performance, jump directly to RPe as our starting frequency.
> > @@ -6839,7 +6839,7 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv)
> >   		freq = max(rps->cur_freq,
> >   			   rps->efficient_freq);
> > -		if (intel_set_rps(dev_priv,
> > +		if (intel_set_rps(i915,
> >   				  clamp(freq,
> >   					rps->min_freq_softlimit,
> >   					rps->max_freq_softlimit)))
> > @@ -6848,26 +6848,26 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv)
> >   	mutex_unlock(&rps->lock);
> >   }
> > -void gen6_rps_idle(struct drm_i915_private *dev_priv)
> > +void gen6_rps_idle(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/* Flush our bottom-half so that it does not race with us
> >   	 * setting the idle frequency and so that it is bounded by
> >   	 * our rpm wakeref. And then disable the interrupts to stop any
> >   	 * futher RPS reclocking whilst we are asleep.
> >   	 */
> > -	gen6_disable_rps_interrupts(dev_priv);
> > +	gen6_disable_rps_interrupts(i915);
> >   	mutex_lock(&rps->lock);
> >   	if (rps->enabled) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -			vlv_set_rps_idle(dev_priv);
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +			vlv_set_rps_idle(i915);
> >   		else
> > -			gen6_set_rps(dev_priv, rps->idle_freq);
> > +			gen6_set_rps(i915, rps->idle_freq);
> >   		rps->last_adj = 0;
> >   		I915_WRITE(GEN6_PMINTRMSK,
> > -			   gen6_sanitize_rps_pm_mask(dev_priv, ~0));
> > +			   gen6_sanitize_rps_pm_mask(i915, ~0));
> >   	}
> >   	mutex_unlock(&rps->lock);
> >   }
> > @@ -6904,9 +6904,9 @@ void gen6_rps_boost(struct i915_request *rq)
> >   	atomic_inc(&rps->boosts);
> >   }
> > -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val)
> > +int intel_set_rps(struct drm_i915_private *i915, u8 val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	int err;
> >   	lockdep_assert_held(&rps->lock);
> > @@ -6918,63 +6918,63 @@ int intel_set_rps(struct drm_i915_private *dev_priv, u8 val)
> >   		return 0;
> >   	}
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
> > -		err = valleyview_set_rps(dev_priv, val);
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
> > +		err = valleyview_set_rps(i915, val);
> >   	else
> > -		err = gen6_set_rps(dev_priv, val);
> > +		err = gen6_set_rps(i915, val);
> >   	return err;
> >   }
> > -static void gen9_disable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen9_disable_rc6(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> >   	I915_WRITE(GEN9_PG_ENABLE, 0);
> >   }
> > -static void gen9_disable_rps(struct drm_i915_private *dev_priv)
> > +static void gen9_disable_rps(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RP_CONTROL, 0);
> >   }
> > -static void gen6_disable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen6_disable_rc6(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> >   }
> > -static void gen6_disable_rps(struct drm_i915_private *dev_priv)
> > +static void gen6_disable_rps(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RPNSWREQ, 1 << 31);
> >   	I915_WRITE(GEN6_RP_CONTROL, 0);
> >   }
> > -static void cherryview_disable_rc6(struct drm_i915_private *dev_priv)
> > +static void cherryview_disable_rc6(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> >   }
> > -static void cherryview_disable_rps(struct drm_i915_private *dev_priv)
> > +static void cherryview_disable_rps(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RP_CONTROL, 0);
> >   }
> > -static void valleyview_disable_rc6(struct drm_i915_private *dev_priv)
> > +static void valleyview_disable_rc6(struct drm_i915_private *i915)
> >   {
> >   	/* We're doing forcewake before Disabling RC6,
> >   	 * This what the BIOS expects when going into suspend */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void valleyview_disable_rps(struct drm_i915_private *dev_priv)
> > +static void valleyview_disable_rps(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(GEN6_RP_CONTROL, 0);
> >   }
> > -static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv)
> > +static bool bxt_check_bios_rc6_setup(struct drm_i915_private *i915)
> >   {
> >   	bool enable_rc6 = true;
> >   	unsigned long rc6_ctx_base;
> > @@ -7000,8 +7000,8 @@ static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv)
> >   	 * for this check.
> >   	 */
> >   	rc6_ctx_base = I915_READ(RC6_CTX_BASE) & RC6_CTX_BASE_MASK;
> > -	if (!((rc6_ctx_base >= dev_priv->dsm_reserved.start) &&
> > -	      (rc6_ctx_base + PAGE_SIZE < dev_priv->dsm_reserved.end))) {
> > +	if (!((rc6_ctx_base >= i915->dsm_reserved.start) &&
> > +	      (rc6_ctx_base + PAGE_SIZE < i915->dsm_reserved.end))) {
> >   		DRM_DEBUG_DRIVER("RC6 Base address not as expected.\n");
> >   		enable_rc6 = false;
> >   	}
> > @@ -7062,14 +7062,14 @@ static bool sanitize_rc6(struct drm_i915_private *i915)
> >   	return info->has_rc6;
> >   }
> > -static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> > +static void gen6_init_rps_frequencies(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/* All of these values are in units of 50MHz */
> >   	/* static values from HW: RP0 > RP1 > RPn (min_freq) */
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		u32 rp_state_cap = I915_READ(BXT_RP_STATE_CAP);
> >   		rps->rp0_freq = (rp_state_cap >> 16) & 0xff;
> >   		rps->rp1_freq = (rp_state_cap >>  8) & 0xff;
> > @@ -7084,11 +7084,11 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> >   	rps->max_freq = rps->rp0_freq;
> >   	rps->efficient_freq = rps->rp1_freq;
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) ||
> > -	    IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915) ||
> > +	    IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> >   		u32 ddcc_status = 0;
> > -		if (sandybridge_pcode_read(dev_priv,
> > +		if (sandybridge_pcode_read(i915,
> >   					   HSW_PCODE_DYNAMIC_DUTY_CYCLE_CONTROL,
> >   					   &ddcc_status, NULL) == 0)
> >   			rps->efficient_freq =
> > @@ -7098,7 +7098,7 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> >   					rps->max_freq);
> >   	}
> > -	if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +	if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> >   		/* Store the frequency values in 16.66 MHZ units, which is
> >   		 * the natural hardware unit for SKL
> >   		 */
> > @@ -7110,45 +7110,45 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void reset_rps(struct drm_i915_private *dev_priv,
> > +static void reset_rps(struct drm_i915_private *i915,
> >   		      int (*set)(struct drm_i915_private *, u8))
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u8 freq = rps->cur_freq;
> >   	/* force a reset */
> >   	rps->power.mode = -1;
> >   	rps->cur_freq = -1;
> > -	if (set(dev_priv, freq))
> > +	if (set(i915, freq))
> >   		DRM_ERROR("Failed to reset RPS to initial values\n");
> >   }
> >   /* See the Gen9_GT_PM_Programming_Guide doc for the below */
> > -static void gen9_enable_rps(struct drm_i915_private *dev_priv)
> > +static void gen9_enable_rps(struct drm_i915_private *i915)
> >   {
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* Program defaults and thresholds for RPS */
> > -	if (IS_GEN(dev_priv, 9))
> > +	if (IS_GEN(i915, 9))
> >   		I915_WRITE(GEN6_RC_VIDEO_FREQ,
> > -			GEN9_FREQUENCY(dev_priv->gt_pm.rps.rp1_freq));
> > +			GEN9_FREQUENCY(i915->gt_pm.rps.rp1_freq));
> >   	/* 1 second timeout*/
> >   	I915_WRITE(GEN6_RP_DOWN_TIMEOUT,
> > -		GT_INTERVAL_FROM_US(dev_priv, 1000000));
> > +		GT_INTERVAL_FROM_US(i915, 1000000));
> >   	I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 0xa);
> >   	/* Leaning on the below call to gen6_set_rps to program/setup the
> >   	 * Up/Down EI & threshold registers, as well as the RP_CONTROL,
> >   	 * RP_INTERRUPT_LIMITS & RPNSWREQ registers */
> > -	reset_rps(dev_priv, gen6_set_rps);
> > +	reset_rps(i915, gen6_set_rps);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen11_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -7160,7 +7160,7 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> >   	 * 1b: Get forcewake during program sequence. Although the driver
> >   	 * hasn't enabled a state yet where we need forcewake, BIOS may have.
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* 2a: Disable RC states. */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > @@ -7171,10 +7171,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> > -	if (HAS_GUC(dev_priv))
> > +	if (HAS_GUC(i915))
> >   		I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA);
> >   	I915_WRITE(GEN6_RC_SLEEP, 0);
> > @@ -7217,10 +7217,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv)
> >   		   GEN9_MEDIA_PG_ENABLE |
> >   		   GEN11_MEDIA_SAMPLER_PG_ENABLE);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen9_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -7231,16 +7231,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> >   	/* 1b: Get forcewake during program sequence. Although the driver
> >   	 * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* 2a: Disable RC states. */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> >   	/* 2b: Program RC6 thresholds.*/
> > -	if (INTEL_GEN(dev_priv) >= 10) {
> > +	if (INTEL_GEN(i915) >= 10) {
> >   		I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 54 << 16 | 85);
> >   		I915_WRITE(GEN10_MEDIA_WAKE_RATE_LIMIT, 150);
> > -	} else if (IS_SKYLAKE(dev_priv)) {
> > +	} else if (IS_SKYLAKE(i915)) {
> >   		/*
> >   		 * WaRsDoubleRc6WrlWithCoarsePowerGating:skl Doubling WRL only
> >   		 * when CPG is enabled
> > @@ -7252,10 +7252,10 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> > -	if (HAS_GUC(dev_priv))
> > +	if (HAS_GUC(i915))
> >   		I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA);
> >   	I915_WRITE(GEN6_RC_SLEEP, 0);
> > @@ -7288,7 +7288,7 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC6_THRESHOLD, 37500); /* 37.5/125ms per EI */
> >   	/* WaRsUseTimeoutMode:cnl (pre-prod) */
> > -	if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_C0))
> > +	if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_C0))
> >   		rc6_mode = GEN7_RC_CTL_TO_MODE;
> >   	else
> >   		rc6_mode = GEN6_RC_CTL_EI_MODE(1);
> > @@ -7302,16 +7302,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv)
> >   	 * 3b: Enable Coarse Power Gating only when RC6 is enabled.
> >   	 * WaRsDisableCoarsePowerGating:skl,cnl - Render/Media PG need to be disabled with RC6.
> >   	 */
> > -	if (NEEDS_WaRsDisableCoarsePowerGating(dev_priv))
> > +	if (NEEDS_WaRsDisableCoarsePowerGating(i915))
> >   		I915_WRITE(GEN9_PG_ENABLE, 0);
> >   	else
> >   		I915_WRITE(GEN9_PG_ENABLE,
> >   			   GEN9_RENDER_PG_ENABLE | GEN9_MEDIA_PG_ENABLE);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen8_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -7321,7 +7321,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> >   	/* 1b: Get forcewake during program sequence. Although the driver
> >   	 * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* 2a: Disable RC states. */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > @@ -7330,7 +7330,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 40 << 16);
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> >   	I915_WRITE(GEN6_RC_SLEEP, 0);
> >   	I915_WRITE(GEN6_RC6_THRESHOLD, 625); /* 800us/1.28 for TO */
> > @@ -7342,14 +7342,14 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv)
> >   		   GEN7_RC_CTL_TO_MODE |
> >   		   GEN6_RC_CTL_RC6_ENABLE);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen8_enable_rps(struct drm_i915_private *dev_priv)
> > +static void gen8_enable_rps(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* 1 Program defaults and thresholds for RPS*/
> >   	I915_WRITE(GEN6_RPNSWREQ,
> > @@ -7380,12 +7380,12 @@ static void gen8_enable_rps(struct drm_i915_private *dev_priv)
> >   		   GEN6_RP_UP_BUSY_AVG |
> >   		   GEN6_RP_DOWN_IDLE_AVG);
> > -	reset_rps(dev_priv, gen6_set_rps);
> > +	reset_rps(i915, gen6_set_rps);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void gen6_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -7402,7 +7402,7 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(GTFIFODBG, gtfifodbg);
> >   	}
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* disable the counters and set deterministic thresholds */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > @@ -7413,12 +7413,12 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000);
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> >   	I915_WRITE(GEN6_RC_SLEEP, 0);
> >   	I915_WRITE(GEN6_RC1e_THRESHOLD, 1000);
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		I915_WRITE(GEN6_RC6_THRESHOLD, 125000);
> >   	else
> >   		I915_WRITE(GEN6_RC6_THRESHOLD, 50000);
> > @@ -7427,9 +7427,9 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> >   	/* We don't use those on Haswell */
> >   	rc6_mask = GEN6_RC_CTL_RC6_ENABLE;
> > -	if (HAS_RC6p(dev_priv))
> > +	if (HAS_RC6p(i915))
> >   		rc6_mask |= GEN6_RC_CTL_RC6p_ENABLE;
> > -	if (HAS_RC6pp(dev_priv))
> > +	if (HAS_RC6pp(i915))
> >   		rc6_mask |= GEN6_RC_CTL_RC6pp_ENABLE;
> >   	I915_WRITE(GEN6_RC_CONTROL,
> >   		   rc6_mask |
> > @@ -7437,24 +7437,24 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv)
> >   		   GEN6_RC_CTL_HW_ENABLE);
> >   	rc6vids = 0;
> > -	ret = sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS,
> > +	ret = sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS,
> >   				     &rc6vids, NULL);
> > -	if (IS_GEN(dev_priv, 6) && ret) {
> > +	if (IS_GEN(i915, 6) && ret) {
> >   		DRM_DEBUG_DRIVER("Couldn't check for BIOS workaround\n");
> > -	} else if (IS_GEN(dev_priv, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) {
> > +	} else if (IS_GEN(i915, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) {
> >   		DRM_DEBUG_DRIVER("You should update your BIOS. Correcting minimum rc6 voltage (%dmV->%dmV)\n",
> >   			  GEN6_DECODE_RC6_VID(rc6vids & 0xff), 450);
> >   		rc6vids &= 0xffff00;
> >   		rc6vids |= GEN6_ENCODE_RC6_VID(450);
> > -		ret = sandybridge_pcode_write(dev_priv, GEN6_PCODE_WRITE_RC6VIDS, rc6vids);
> > +		ret = sandybridge_pcode_write(i915, GEN6_PCODE_WRITE_RC6VIDS, rc6vids);
> >   		if (ret)
> >   			DRM_ERROR("Couldn't fix incorrect rc6 voltage\n");
> >   	}
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen6_enable_rps(struct drm_i915_private *dev_priv)
> > +static void gen6_enable_rps(struct drm_i915_private *i915)
> >   {
> >   	/* Here begins a magic sequence of register writes to enable
> >   	 * auto-downclocking.
> > @@ -7462,20 +7462,20 @@ static void gen6_enable_rps(struct drm_i915_private *dev_priv)
> >   	 * Perhaps there might be some value in exposing these to
> >   	 * userspace...
> >   	 */
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* Power down if completely idle for over 50ms */
> >   	I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 50000);
> >   	I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 10);
> > -	reset_rps(dev_priv, gen6_set_rps);
> > +	reset_rps(i915, gen6_set_rps);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> > +static void gen6_update_ring_freq(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	const int min_freq = 15;
> >   	const int scaling_factor = 180;
> >   	unsigned int gpu_freq;
> > @@ -7509,7 +7509,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> >   	min_gpu_freq = rps->min_freq;
> >   	max_gpu_freq = rps->max_freq;
> > -	if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +	if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> >   		/* Convert GT frequency to 50 HZ units */
> >   		min_gpu_freq /= GEN9_FREQ_SCALER;
> >   		max_gpu_freq /= GEN9_FREQ_SCALER;
> > @@ -7524,16 +7524,16 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> >   		const int diff = max_gpu_freq - gpu_freq;
> >   		unsigned int ia_freq = 0, ring_freq = 0;
> > -		if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) {
> > +		if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) {
> >   			/*
> >   			 * ring_freq = 2 * GT. ring_freq is in 100MHz units
> >   			 * No floor required for ring frequency on SKL.
> >   			 */
> >   			ring_freq = gpu_freq;
> > -		} else if (INTEL_GEN(dev_priv) >= 8) {
> > +		} else if (INTEL_GEN(i915) >= 8) {
> >   			/* max(2 * GT, DDR). NB: GT is 50MHz units */
> >   			ring_freq = max(min_ring_freq, gpu_freq);
> > -		} else if (IS_HASWELL(dev_priv)) {
> > +		} else if (IS_HASWELL(i915)) {
> >   			ring_freq = mult_frac(gpu_freq, 5, 4);
> >   			ring_freq = max(min_ring_freq, ring_freq);
> >   			/* leave ia_freq as the default, chosen by cpufreq */
> > @@ -7552,7 +7552,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> >   			ia_freq = DIV_ROUND_CLOSEST(ia_freq, 100);
> >   		}
> > -		sandybridge_pcode_write(dev_priv,
> > +		sandybridge_pcode_write(i915,
> >   					GEN6_PCODE_WRITE_MIN_FREQ_TABLE,
> >   					ia_freq << GEN6_PCODE_FREQ_IA_RATIO_SHIFT |
> >   					ring_freq << GEN6_PCODE_FREQ_RING_RATIO_SHIFT |
> > @@ -7560,13 +7560,13 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv)
> > +static int cherryview_rps_max_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rp0;
> > -	val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE);
> > +	val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE);
> > -	switch (RUNTIME_INFO(dev_priv)->sseu.eu_total) {
> > +	switch (RUNTIME_INFO(i915)->sseu.eu_total) {
> >   	case 8:
> >   		/* (2 * 4) config */
> >   		rp0 = (val >> FB_GFX_FMAX_AT_VMAX_2SS4EU_FUSE_SHIFT);
> > @@ -7588,53 +7588,53 @@ static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv)
> >   	return rp0;
> >   }
> > -static int cherryview_rps_rpe_freq(struct drm_i915_private *dev_priv)
> > +static int cherryview_rps_rpe_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rpe;
> > -	val = vlv_punit_read(dev_priv, PUNIT_GPU_DUTYCYCLE_REG);
> > +	val = vlv_punit_read(i915, PUNIT_GPU_DUTYCYCLE_REG);
> >   	rpe = (val >> PUNIT_GPU_DUTYCYCLE_RPE_FREQ_SHIFT) & PUNIT_GPU_DUTYCYCLE_RPE_FREQ_MASK;
> >   	return rpe;
> >   }
> > -static int cherryview_rps_guar_freq(struct drm_i915_private *dev_priv)
> > +static int cherryview_rps_guar_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rp1;
> > -	val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE);
> > +	val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE);
> >   	rp1 = (val & FB_GFX_FREQ_FUSE_MASK);
> >   	return rp1;
> >   }
> > -static u32 cherryview_rps_min_freq(struct drm_i915_private *dev_priv)
> > +static u32 cherryview_rps_min_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rpn;
> > -	val = vlv_punit_read(dev_priv, FB_GFX_FMIN_AT_VMIN_FUSE);
> > +	val = vlv_punit_read(i915, FB_GFX_FMIN_AT_VMIN_FUSE);
> >   	rpn = ((val >> FB_GFX_FMIN_AT_VMIN_FUSE_SHIFT) &
> >   		       FB_GFX_FREQ_FUSE_MASK);
> >   	return rpn;
> >   }
> > -static int valleyview_rps_guar_freq(struct drm_i915_private *dev_priv)
> > +static int valleyview_rps_guar_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rp1;
> > -	val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE);
> > +	val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE);
> >   	rp1 = (val & FB_GFX_FGUARANTEED_FREQ_FUSE_MASK) >> FB_GFX_FGUARANTEED_FREQ_FUSE_SHIFT;
> >   	return rp1;
> >   }
> > -static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv)
> > +static int valleyview_rps_max_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rp0;
> > -	val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE);
> > +	val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE);
> >   	rp0 = (val & FB_GFX_MAX_FREQ_FUSE_MASK) >> FB_GFX_MAX_FREQ_FUSE_SHIFT;
> >   	/* Clamp to max */
> > @@ -7643,23 +7643,23 @@ static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv)
> >   	return rp0;
> >   }
> > -static int valleyview_rps_rpe_freq(struct drm_i915_private *dev_priv)
> > +static int valleyview_rps_rpe_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val, rpe;
> > -	val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_LO);
> > +	val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_LO);
> >   	rpe = (val & FB_FMAX_VMIN_FREQ_LO_MASK) >> FB_FMAX_VMIN_FREQ_LO_SHIFT;
> > -	val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_HI);
> > +	val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_HI);
> >   	rpe |= (val & FB_FMAX_VMIN_FREQ_HI_MASK) << 5;
> >   	return rpe;
> >   }
> > -static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv)
> > +static int valleyview_rps_min_freq(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_LFM) & 0xff;
> > +	val = vlv_punit_read(i915, PUNIT_REG_GPU_LFM) & 0xff;
> >   	/*
> >   	 * According to the BYT Punit GPU turbo HAS 1.1.6.3 the minimum value
> >   	 * for the minimum frequency in GPLL mode is 0xc1. Contrary to this on
> > @@ -7671,24 +7671,24 @@ static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv)
> >   }
> >   /* Check that the pctx buffer wasn't move under us. */
> > -static void valleyview_check_pctx(struct drm_i915_private *dev_priv)
> > +static void valleyview_check_pctx(struct drm_i915_private *i915)
> >   {
> >   	unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095;
> > -	WARN_ON(pctx_addr != dev_priv->dsm.start +
> > -			     dev_priv->vlv_pctx->stolen->start);
> > +	WARN_ON(pctx_addr != i915->dsm.start +
> > +			     i915->vlv_pctx->stolen->start);
> >   }
> >   /* Check that the pcbr address is not empty. */
> > -static void cherryview_check_pctx(struct drm_i915_private *dev_priv)
> > +static void cherryview_check_pctx(struct drm_i915_private *i915)
> >   {
> >   	unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095;
> >   	WARN_ON((pctx_addr >> VLV_PCBR_ADDR_SHIFT) == 0);
> >   }
> > -static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> > +static void cherryview_setup_pctx(struct drm_i915_private *i915)
> >   {
> >   	resource_size_t pctx_paddr, paddr;
> >   	resource_size_t pctx_size = 32*1024;
> > @@ -7697,7 +7697,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> >   	pcbr = I915_READ(VLV_PCBR);
> >   	if ((pcbr >> VLV_PCBR_ADDR_SHIFT) == 0) {
> >   		DRM_DEBUG_DRIVER("BIOS didn't set up PCBR, fixing up\n");
> > -		paddr = dev_priv->dsm.end + 1 - pctx_size;
> > +		paddr = i915->dsm.end + 1 - pctx_size;
> >   		GEM_BUG_ON(paddr > U32_MAX);
> >   		pctx_paddr = (paddr & (~4095));
> > @@ -7707,7 +7707,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR));
> >   }
> > -static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> > +static void valleyview_setup_pctx(struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_gem_object *pctx;
> >   	resource_size_t pctx_paddr;
> > @@ -7719,8 +7719,8 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> >   		/* BIOS set it up already, grab the pre-alloc'd space */
> >   		resource_size_t pcbr_offset;
> > -		pcbr_offset = (pcbr & (~4095)) - dev_priv->dsm.start;
> > -		pctx = i915_gem_object_create_stolen_for_preallocated(dev_priv,
> > +		pcbr_offset = (pcbr & (~4095)) - i915->dsm.start;
> > +		pctx = i915_gem_object_create_stolen_for_preallocated(i915,
> >   								      pcbr_offset,
> >   								      I915_GTT_OFFSET_NONE,
> >   								      pctx_size);
> > @@ -7737,148 +7737,148 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv)
> >   	 * overlap with other ranges, such as the frame buffer, protected
> >   	 * memory, or any other relevant ranges.
> >   	 */
> > -	pctx = i915_gem_object_create_stolen(dev_priv, pctx_size);
> > +	pctx = i915_gem_object_create_stolen(i915, pctx_size);
> >   	if (!pctx) {
> >   		DRM_DEBUG("not enough stolen space for PCTX, disabling\n");
> >   		goto out;
> >   	}
> >   	GEM_BUG_ON(range_overflows_t(u64,
> > -				     dev_priv->dsm.start,
> > +				     i915->dsm.start,
> >   				     pctx->stolen->start,
> >   				     U32_MAX));
> > -	pctx_paddr = dev_priv->dsm.start + pctx->stolen->start;
> > +	pctx_paddr = i915->dsm.start + pctx->stolen->start;
> >   	I915_WRITE(VLV_PCBR, pctx_paddr);
> >   out:
> >   	DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR));
> > -	dev_priv->vlv_pctx = pctx;
> > +	i915->vlv_pctx = pctx;
> >   }
> > -static void valleyview_cleanup_pctx(struct drm_i915_private *dev_priv)
> > +static void valleyview_cleanup_pctx(struct drm_i915_private *i915)
> >   {
> >   	struct drm_i915_gem_object *pctx;
> > -	pctx = fetch_and_zero(&dev_priv->vlv_pctx);
> > +	pctx = fetch_and_zero(&i915->vlv_pctx);
> >   	if (pctx)
> >   		i915_gem_object_put(pctx);
> >   }
> > -static void vlv_init_gpll_ref_freq(struct drm_i915_private *dev_priv)
> > +static void vlv_init_gpll_ref_freq(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->gt_pm.rps.gpll_ref_freq =
> > -		vlv_get_cck_clock(dev_priv, "GPLL ref",
> > +	i915->gt_pm.rps.gpll_ref_freq =
> > +		vlv_get_cck_clock(i915, "GPLL ref",
> >   				  CCK_GPLL_CLOCK_CONTROL,
> > -				  dev_priv->czclk_freq);
> > +				  i915->czclk_freq);
> >   	DRM_DEBUG_DRIVER("GPLL reference freq: %d kHz\n",
> > -			 dev_priv->gt_pm.rps.gpll_ref_freq);
> > +			 i915->gt_pm.rps.gpll_ref_freq);
> >   }
> > -static void valleyview_init_gt_powersave(struct drm_i915_private *dev_priv)
> > +static void valleyview_init_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 val;
> > -	valleyview_setup_pctx(dev_priv);
> > +	valleyview_setup_pctx(i915);
> > -	vlv_iosf_sb_get(dev_priv,
> > +	vlv_iosf_sb_get(i915,
> >   			BIT(VLV_IOSF_SB_PUNIT) |
> >   			BIT(VLV_IOSF_SB_NC) |
> >   			BIT(VLV_IOSF_SB_CCK));
> > -	vlv_init_gpll_ref_freq(dev_priv);
> > +	vlv_init_gpll_ref_freq(i915);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> > +	val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> >   	switch ((val >> 6) & 3) {
> >   	case 0:
> >   	case 1:
> > -		dev_priv->mem_freq = 800;
> > +		i915->mem_freq = 800;
> >   		break;
> >   	case 2:
> > -		dev_priv->mem_freq = 1066;
> > +		i915->mem_freq = 1066;
> >   		break;
> >   	case 3:
> > -		dev_priv->mem_freq = 1333;
> > +		i915->mem_freq = 1333;
> >   		break;
> >   	}
> > -	DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq);
> > +	DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq);
> > -	rps->max_freq = valleyview_rps_max_freq(dev_priv);
> > +	rps->max_freq = valleyview_rps_max_freq(i915);
> >   	rps->rp0_freq = rps->max_freq;
> >   	DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->max_freq),
> > +			 intel_gpu_freq(i915, rps->max_freq),
> >   			 rps->max_freq);
> > -	rps->efficient_freq = valleyview_rps_rpe_freq(dev_priv);
> > +	rps->efficient_freq = valleyview_rps_rpe_freq(i915);
> >   	DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->efficient_freq),
> > +			 intel_gpu_freq(i915, rps->efficient_freq),
> >   			 rps->efficient_freq);
> > -	rps->rp1_freq = valleyview_rps_guar_freq(dev_priv);
> > +	rps->rp1_freq = valleyview_rps_guar_freq(i915);
> >   	DRM_DEBUG_DRIVER("RP1(Guar Freq) GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->rp1_freq),
> > +			 intel_gpu_freq(i915, rps->rp1_freq),
> >   			 rps->rp1_freq);
> > -	rps->min_freq = valleyview_rps_min_freq(dev_priv);
> > +	rps->min_freq = valleyview_rps_min_freq(i915);
> >   	DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->min_freq),
> > +			 intel_gpu_freq(i915, rps->min_freq),
> >   			 rps->min_freq);
> > -	vlv_iosf_sb_put(dev_priv,
> > +	vlv_iosf_sb_put(i915,
> >   			BIT(VLV_IOSF_SB_PUNIT) |
> >   			BIT(VLV_IOSF_SB_NC) |
> >   			BIT(VLV_IOSF_SB_CCK));
> >   }
> > -static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv)
> > +static void cherryview_init_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	u32 val;
> > -	cherryview_setup_pctx(dev_priv);
> > +	cherryview_setup_pctx(i915);
> > -	vlv_iosf_sb_get(dev_priv,
> > +	vlv_iosf_sb_get(i915,
> >   			BIT(VLV_IOSF_SB_PUNIT) |
> >   			BIT(VLV_IOSF_SB_NC) |
> >   			BIT(VLV_IOSF_SB_CCK));
> > -	vlv_init_gpll_ref_freq(dev_priv);
> > +	vlv_init_gpll_ref_freq(i915);
> > -	val = vlv_cck_read(dev_priv, CCK_FUSE_REG);
> > +	val = vlv_cck_read(i915, CCK_FUSE_REG);
> >   	switch ((val >> 2) & 0x7) {
> >   	case 3:
> > -		dev_priv->mem_freq = 2000;
> > +		i915->mem_freq = 2000;
> >   		break;
> >   	default:
> > -		dev_priv->mem_freq = 1600;
> > +		i915->mem_freq = 1600;
> >   		break;
> >   	}
> > -	DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq);
> > +	DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq);
> > -	rps->max_freq = cherryview_rps_max_freq(dev_priv);
> > +	rps->max_freq = cherryview_rps_max_freq(i915);
> >   	rps->rp0_freq = rps->max_freq;
> >   	DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->max_freq),
> > +			 intel_gpu_freq(i915, rps->max_freq),
> >   			 rps->max_freq);
> > -	rps->efficient_freq = cherryview_rps_rpe_freq(dev_priv);
> > +	rps->efficient_freq = cherryview_rps_rpe_freq(i915);
> >   	DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->efficient_freq),
> > +			 intel_gpu_freq(i915, rps->efficient_freq),
> >   			 rps->efficient_freq);
> > -	rps->rp1_freq = cherryview_rps_guar_freq(dev_priv);
> > +	rps->rp1_freq = cherryview_rps_guar_freq(i915);
> >   	DRM_DEBUG_DRIVER("RP1(Guar) GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->rp1_freq),
> > +			 intel_gpu_freq(i915, rps->rp1_freq),
> >   			 rps->rp1_freq);
> > -	rps->min_freq = cherryview_rps_min_freq(dev_priv);
> > +	rps->min_freq = cherryview_rps_min_freq(i915);
> >   	DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n",
> > -			 intel_gpu_freq(dev_priv, rps->min_freq),
> > +			 intel_gpu_freq(i915, rps->min_freq),
> >   			 rps->min_freq);
> > -	vlv_iosf_sb_put(dev_priv,
> > +	vlv_iosf_sb_put(i915,
> >   			BIT(VLV_IOSF_SB_PUNIT) |
> >   			BIT(VLV_IOSF_SB_NC) |
> >   			BIT(VLV_IOSF_SB_CCK));
> > @@ -7888,12 +7888,12 @@ static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv)
> >   		  "Odd GPU freq values\n");
> >   }
> > -static void valleyview_cleanup_gt_powersave(struct drm_i915_private *dev_priv)
> > +static void valleyview_cleanup_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	valleyview_cleanup_pctx(dev_priv);
> > +	valleyview_cleanup_pctx(i915);
> >   }
> > -static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void cherryview_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> > @@ -7907,11 +7907,11 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(GTFIFODBG, gtfifodbg);
> >   	}
> > -	cherryview_check_pctx(dev_priv);
> > +	cherryview_check_pctx(i915);
> >   	/* 1a & 1b: Get forcewake during program sequence. Although the driver
> >   	 * hasn't enabled a state yet where we need forcewake, BIOS may have.*/
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/*  Disable RC states. */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > @@ -7921,7 +7921,7 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> >   	I915_WRITE(GEN6_RC_SLEEP, 0);
> > @@ -7943,14 +7943,14 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv)
> >   		rc6_mode = GEN7_RC_CTL_TO_MODE;
> >   	I915_WRITE(GEN6_RC_CONTROL, rc6_mode);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> > +static void cherryview_enable_rps(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/* 1: Program defaults and thresholds for RPS*/
> >   	I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000);
> > @@ -7970,14 +7970,14 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> >   		   GEN6_RP_DOWN_IDLE_AVG);
> >   	/* Setting Fixed Bias */
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> >   	val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | CHV_BIAS_CPU_50_SOC_50;
> > -	vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val);
> > +	vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> > +	val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   	/* RPS code assumes GPLL is used */
> >   	WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n");
> > @@ -7985,18 +7985,18 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE));
> >   	DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val);
> > -	reset_rps(dev_priv, valleyview_set_rps);
> > +	reset_rps(i915, valleyview_set_rps);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void valleyview_enable_rc6(struct drm_i915_private *i915)
> >   {
> >   	struct intel_engine_cs *engine;
> >   	enum intel_engine_id id;
> >   	u32 gtfifodbg;
> > -	valleyview_check_pctx(dev_priv);
> > +	valleyview_check_pctx(i915);
> >   	gtfifodbg = I915_READ(GTFIFODBG);
> >   	if (gtfifodbg) {
> > @@ -8005,7 +8005,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(GTFIFODBG, gtfifodbg);
> >   	}
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	/*  Disable RC states. */
> >   	I915_WRITE(GEN6_RC_CONTROL, 0);
> > @@ -8014,7 +8014,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000);
> >   	I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25);
> > -	for_each_engine(engine, dev_priv, id)
> > +	for_each_engine(engine, i915, id)
> >   		I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10);
> >   	I915_WRITE(GEN6_RC6_THRESHOLD, 0x557);
> > @@ -8030,14 +8030,14 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN6_RC_CONTROL,
> >   		   GEN7_RC_CTL_TO_MODE | VLV_RC_CTL_CTX_RST_PARALLEL);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> > -static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> > +static void valleyview_enable_rps(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> >   	I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000);
> >   	I915_WRITE(GEN6_RP_UP_THRESHOLD, 59400);
> > @@ -8055,15 +8055,15 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> >   		   GEN6_RP_UP_BUSY_AVG |
> >   		   GEN6_RP_DOWN_IDLE_CONT);
> > -	vlv_punit_get(dev_priv);
> > +	vlv_punit_get(i915);
> >   	/* Setting Fixed Bias */
> >   	val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | VLV_BIAS_CPU_125_SOC_875;
> > -	vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val);
> > +	vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val);
> > -	val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS);
> > +	val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
> > -	vlv_punit_put(dev_priv);
> > +	vlv_punit_put(i915);
> >   	/* RPS code assumes GPLL is used */
> >   	WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n");
> > @@ -8071,9 +8071,9 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE));
> >   	DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val);
> > -	reset_rps(dev_priv, valleyview_set_rps);
> > +	reset_rps(i915, valleyview_set_rps);
> > -	intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +	intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   }
> >   static unsigned long intel_pxfreq(u32 vidfreq)
> > @@ -8105,7 +8105,7 @@ static const struct cparams {
> >   	{ 0, 800, 231, 23784 },
> >   };
> > -static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> > +static unsigned long __i915_chipset_val(struct drm_i915_private *i915)
> >   {
> >   	u64 total_count, diff, ret;
> >   	u32 count1, count2, count3, m = 0, c = 0;
> > @@ -8114,7 +8114,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> >   	lockdep_assert_held(&mchdev_lock);
> > -	diff1 = now - dev_priv->ips.last_time1;
> > +	diff1 = now - i915->ips.last_time1;
> >   	/* Prevent division-by-zero if we are asking too fast.
> >   	 * Also, we don't get interesting results if we are polling
> > @@ -8122,7 +8122,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> >   	 * in such cases.
> >   	 */
> >   	if (diff1 <= 10)
> > -		return dev_priv->ips.chipset_power;
> > +		return i915->ips.chipset_power;
> >   	count1 = I915_READ(DMIEC);
> >   	count2 = I915_READ(DDREC);
> > @@ -8131,16 +8131,16 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> >   	total_count = count1 + count2 + count3;
> >   	/* FIXME: handle per-counter overflow */
> > -	if (total_count < dev_priv->ips.last_count1) {
> > -		diff = ~0UL - dev_priv->ips.last_count1;
> > +	if (total_count < i915->ips.last_count1) {
> > +		diff = ~0UL - i915->ips.last_count1;
> >   		diff += total_count;
> >   	} else {
> > -		diff = total_count - dev_priv->ips.last_count1;
> > +		diff = total_count - i915->ips.last_count1;
> >   	}
> >   	for (i = 0; i < ARRAY_SIZE(cparams); i++) {
> > -		if (cparams[i].i == dev_priv->ips.c_m &&
> > -		    cparams[i].t == dev_priv->ips.r_t) {
> > +		if (cparams[i].i == i915->ips.c_m &&
> > +		    cparams[i].t == i915->ips.r_t) {
> >   			m = cparams[i].m;
> >   			c = cparams[i].c;
> >   			break;
> > @@ -8151,25 +8151,25 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv)
> >   	ret = ((m * diff) + c);
> >   	ret = div_u64(ret, 10);
> > -	dev_priv->ips.last_count1 = total_count;
> > -	dev_priv->ips.last_time1 = now;
> > +	i915->ips.last_count1 = total_count;
> > +	i915->ips.last_time1 = now;
> > -	dev_priv->ips.chipset_power = ret;
> > +	i915->ips.chipset_power = ret;
> >   	return ret;
> >   }
> > -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv)
> > +unsigned long i915_chipset_val(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref;
> >   	unsigned long val = 0;
> > -	if (!IS_GEN(dev_priv, 5))
> > +	if (!IS_GEN(i915, 5))
> >   		return 0;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		spin_lock_irq(&mchdev_lock);
> > -		val = __i915_chipset_val(dev_priv);
> > +		val = __i915_chipset_val(i915);
> >   		spin_unlock_irq(&mchdev_lock);
> >   	}
> > @@ -8202,18 +8202,18 @@ static int _pxvid_to_vd(u8 pxvid)
> >   	return (pxvid + 2) * 125;
> >   }
> > -static u32 pvid_to_extvid(struct drm_i915_private *dev_priv, u8 pxvid)
> > +static u32 pvid_to_extvid(struct drm_i915_private *i915, u8 pxvid)
> >   {
> >   	const int vd = _pxvid_to_vd(pxvid);
> >   	const int vm = vd - 1125;
> > -	if (INTEL_INFO(dev_priv)->is_mobile)
> > +	if (INTEL_INFO(i915)->is_mobile)
> >   		return vm > 0 ? vm : 0;
> >   	return vd;
> >   }
> > -static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
> > +static void __i915_update_gfx_val(struct drm_i915_private *i915)
> >   {
> >   	u64 now, diff, diffms;
> >   	u32 count;
> > @@ -8221,7 +8221,7 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
> >   	lockdep_assert_held(&mchdev_lock);
> >   	now = ktime_get_raw_ns();
> > -	diffms = now - dev_priv->ips.last_time2;
> > +	diffms = now - i915->ips.last_time2;
> >   	do_div(diffms, NSEC_PER_MSEC);
> >   	/* Don't divide by 0 */
> > @@ -8230,50 +8230,50 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv)
> >   	count = I915_READ(GFXEC);
> > -	if (count < dev_priv->ips.last_count2) {
> > -		diff = ~0UL - dev_priv->ips.last_count2;
> > +	if (count < i915->ips.last_count2) {
> > +		diff = ~0UL - i915->ips.last_count2;
> >   		diff += count;
> >   	} else {
> > -		diff = count - dev_priv->ips.last_count2;
> > +		diff = count - i915->ips.last_count2;
> >   	}
> > -	dev_priv->ips.last_count2 = count;
> > -	dev_priv->ips.last_time2 = now;
> > +	i915->ips.last_count2 = count;
> > +	i915->ips.last_time2 = now;
> >   	/* More magic constants... */
> >   	diff = diff * 1181;
> >   	diff = div_u64(diff, diffms * 10);
> > -	dev_priv->ips.gfx_power = diff;
> > +	i915->ips.gfx_power = diff;
> >   }
> > -void i915_update_gfx_val(struct drm_i915_private *dev_priv)
> > +void i915_update_gfx_val(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref;
> > -	if (!IS_GEN(dev_priv, 5))
> > +	if (!IS_GEN(i915, 5))
> >   		return;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		spin_lock_irq(&mchdev_lock);
> > -		__i915_update_gfx_val(dev_priv);
> > +		__i915_update_gfx_val(i915);
> >   		spin_unlock_irq(&mchdev_lock);
> >   	}
> >   }
> > -static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv)
> > +static unsigned long __i915_gfx_val(struct drm_i915_private *i915)
> >   {
> >   	unsigned long t, corr, state1, corr2, state2;
> >   	u32 pxvid, ext_v;
> >   	lockdep_assert_held(&mchdev_lock);
> > -	pxvid = I915_READ(PXVFREQ(dev_priv->gt_pm.rps.cur_freq));
> > +	pxvid = I915_READ(PXVFREQ(i915->gt_pm.rps.cur_freq));
> >   	pxvid = (pxvid >> 24) & 0x7f;
> > -	ext_v = pvid_to_extvid(dev_priv, pxvid);
> > +	ext_v = pvid_to_extvid(i915, pxvid);
> >   	state1 = ext_v;
> > -	t = i915_mch_val(dev_priv);
> > +	t = i915_mch_val(i915);
> >   	/* Revel in the empirically derived constants */
> > @@ -8287,27 +8287,27 @@ static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv)
> >   	corr = corr * ((150142 * state1) / 10000 - 78642);
> >   	corr /= 100000;
> > -	corr2 = (corr * dev_priv->ips.corr);
> > +	corr2 = (corr * i915->ips.corr);
> >   	state2 = (corr2 * state1) / 10000;
> >   	state2 /= 100; /* convert to mW */
> > -	__i915_update_gfx_val(dev_priv);
> > +	__i915_update_gfx_val(i915);
> > -	return dev_priv->ips.gfx_power + state2;
> > +	return i915->ips.gfx_power + state2;
> >   }
> > -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv)
> > +unsigned long i915_gfx_val(struct drm_i915_private *i915)
> >   {
> >   	intel_wakeref_t wakeref;
> >   	unsigned long val = 0;
> > -	if (!IS_GEN(dev_priv, 5))
> > +	if (!IS_GEN(i915, 5))
> >   		return 0;
> > -	with_intel_runtime_pm(dev_priv, wakeref) {
> > +	with_intel_runtime_pm(i915, wakeref) {
> >   		spin_lock_irq(&mchdev_lock);
> > -		val = __i915_gfx_val(dev_priv);
> > +		val = __i915_gfx_val(i915);
> >   		spin_unlock_irq(&mchdev_lock);
> >   	}
> > @@ -8471,11 +8471,11 @@ ips_ping_for_i915_load(void)
> >   	}
> >   }
> > -void intel_gpu_ips_init(struct drm_i915_private *dev_priv)
> > +void intel_gpu_ips_init(struct drm_i915_private *i915)
> >   {
> >   	/* We only register the i915 ips part with intel-ips once everything is
> >   	 * set up, to avoid intel-ips sneaking in and reading bogus values. */
> > -	rcu_assign_pointer(i915_mch_dev, dev_priv);
> > +	rcu_assign_pointer(i915_mch_dev, i915);
> >   	ips_ping_for_i915_load();
> >   }
> > @@ -8485,7 +8485,7 @@ void intel_gpu_ips_teardown(void)
> >   	rcu_assign_pointer(i915_mch_dev, NULL);
> >   }
> > -static void intel_init_emon(struct drm_i915_private *dev_priv)
> > +static void intel_init_emon(struct drm_i915_private *i915)
> >   {
> >   	u32 lcfuse;
> >   	u8 pxw[16];
> > @@ -8552,40 +8552,40 @@ static void intel_init_emon(struct drm_i915_private *dev_priv)
> >   	lcfuse = I915_READ(LCFUSE02);
> > -	dev_priv->ips.corr = (lcfuse & LCFUSE_HIV_MASK);
> > +	i915->ips.corr = (lcfuse & LCFUSE_HIV_MASK);
> >   }
> > -void intel_init_gt_powersave(struct drm_i915_private *dev_priv)
> > +void intel_init_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/*
> >   	 * RPM depends on RC6 to save restore the GT HW context, so make RC6 a
> >   	 * requirement.
> >   	 */
> > -	if (!sanitize_rc6(dev_priv)) {
> > +	if (!sanitize_rc6(i915)) {
> >   		DRM_INFO("RC6 disabled, disabling runtime PM support\n");
> > -		pm_runtime_get(&dev_priv->drm.pdev->dev);
> > +		pm_runtime_get(&i915->drm.pdev->dev);
> >   	}
> >   	/* Initialize RPS limits (for userspace) */
> > -	if (IS_CHERRYVIEW(dev_priv))
> > -		cherryview_init_gt_powersave(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		valleyview_init_gt_powersave(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > -		gen6_init_rps_frequencies(dev_priv);
> > +	if (IS_CHERRYVIEW(i915))
> > +		cherryview_init_gt_powersave(i915);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		valleyview_init_gt_powersave(i915);
> > +	else if (INTEL_GEN(i915) >= 6)
> > +		gen6_init_rps_frequencies(i915);
> >   	/* Derive initial user preferences/limits from the hardware limits */
> >   	rps->max_freq_softlimit = rps->max_freq;
> >   	rps->min_freq_softlimit = rps->min_freq;
> >   	/* After setting max-softlimit, find the overclock max freq */
> > -	if (IS_GEN(dev_priv, 6) ||
> > -	    IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv)) {
> > +	if (IS_GEN(i915, 6) ||
> > +	    IS_IVYBRIDGE(i915) || IS_HASWELL(i915)) {
> >   		u32 params = 0;
> > -		sandybridge_pcode_read(dev_priv, GEN6_READ_OC_PARAMS,
> > +		sandybridge_pcode_read(i915, GEN6_READ_OC_PARAMS,
> >   				       &params, NULL);
> >   		if (params & BIT(31)) { /* OC supported */
> >   			DRM_DEBUG_DRIVER("Overclocking supported, max: %dMHz, overclock: %dMHz\n",
> > @@ -8601,25 +8601,25 @@ void intel_init_gt_powersave(struct drm_i915_private *dev_priv)
> >   	rps->cur_freq = rps->idle_freq;
> >   }
> > -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv)
> > +void intel_cleanup_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	if (IS_VALLEYVIEW(dev_priv))
> > -		valleyview_cleanup_gt_powersave(dev_priv);
> > +	if (IS_VALLEYVIEW(i915))
> > +		valleyview_cleanup_gt_powersave(i915);
> > -	if (!HAS_RC6(dev_priv))
> > -		pm_runtime_put(&dev_priv->drm.pdev->dev);
> > +	if (!HAS_RC6(i915))
> > +		pm_runtime_put(&i915->drm.pdev->dev);
> >   }
> > -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv)
> > +void intel_sanitize_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->gt_pm.rps.enabled = true; /* force RPS disabling */
> > -	dev_priv->gt_pm.rc6.enabled = true; /* force RC6 disabling */
> > -	intel_disable_gt_powersave(dev_priv);
> > +	i915->gt_pm.rps.enabled = true; /* force RPS disabling */
> > +	i915->gt_pm.rc6.enabled = true; /* force RC6 disabling */
> > +	intel_disable_gt_powersave(i915);
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > -		gen11_reset_rps_interrupts(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > -		gen6_reset_rps_interrupts(dev_priv);
> > +	if (INTEL_GEN(i915) >= 11)
> > +		gen11_reset_rps_interrupts(i915);
> > +	else if (INTEL_GEN(i915) >= 6)
> > +		gen6_reset_rps_interrupts(i915);
> >   }
> >   static inline void intel_disable_llc_pstate(struct drm_i915_private *i915)
> > @@ -8634,56 +8634,56 @@ static inline void intel_disable_llc_pstate(struct drm_i915_private *i915)
> >   	i915->gt_pm.llc_pstate.enabled = false;
> >   }
> > -static void intel_disable_rc6(struct drm_i915_private *dev_priv)
> > +static void intel_disable_rc6(struct drm_i915_private *i915)
> >   {
> > -	lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> > +	lockdep_assert_held(&i915->gt_pm.rps.lock);
> > -	if (!dev_priv->gt_pm.rc6.enabled)
> > +	if (!i915->gt_pm.rc6.enabled)
> >   		return;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		gen9_disable_rc6(dev_priv);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		cherryview_disable_rc6(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		valleyview_disable_rc6(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > -		gen6_disable_rc6(dev_priv);
> > +	if (INTEL_GEN(i915) >= 9)
> > +		gen9_disable_rc6(i915);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		cherryview_disable_rc6(i915);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		valleyview_disable_rc6(i915);
> > +	else if (INTEL_GEN(i915) >= 6)
> > +		gen6_disable_rc6(i915);
> > -	dev_priv->gt_pm.rc6.enabled = false;
> > +	i915->gt_pm.rc6.enabled = false;
> >   }
> > -static void intel_disable_rps(struct drm_i915_private *dev_priv)
> > +static void intel_disable_rps(struct drm_i915_private *i915)
> >   {
> > -	lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> > +	lockdep_assert_held(&i915->gt_pm.rps.lock);
> > -	if (!dev_priv->gt_pm.rps.enabled)
> > +	if (!i915->gt_pm.rps.enabled)
> >   		return;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		gen9_disable_rps(dev_priv);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		cherryview_disable_rps(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		valleyview_disable_rps(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > -		gen6_disable_rps(dev_priv);
> > -	else if (IS_IRONLAKE_M(dev_priv))
> > -		ironlake_disable_drps(dev_priv);
> > +	if (INTEL_GEN(i915) >= 9)
> > +		gen9_disable_rps(i915);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		cherryview_disable_rps(i915);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		valleyview_disable_rps(i915);
> > +	else if (INTEL_GEN(i915) >= 6)
> > +		gen6_disable_rps(i915);
> > +	else if (IS_IRONLAKE_M(i915))
> > +		ironlake_disable_drps(i915);
> > -	dev_priv->gt_pm.rps.enabled = false;
> > +	i915->gt_pm.rps.enabled = false;
> >   }
> > -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv)
> > +void intel_disable_gt_powersave(struct drm_i915_private *i915)
> >   {
> > -	mutex_lock(&dev_priv->gt_pm.rps.lock);
> > +	mutex_lock(&i915->gt_pm.rps.lock);
> > -	intel_disable_rc6(dev_priv);
> > -	intel_disable_rps(dev_priv);
> > -	if (HAS_LLC(dev_priv))
> > -		intel_disable_llc_pstate(dev_priv);
> > +	intel_disable_rc6(i915);
> > +	intel_disable_rps(i915);
> > +	if (HAS_LLC(i915))
> > +		intel_disable_llc_pstate(i915);
> > -	mutex_unlock(&dev_priv->gt_pm.rps.lock);
> > +	mutex_unlock(&i915->gt_pm.rps.lock);
> >   }
> >   static inline void intel_enable_llc_pstate(struct drm_i915_private *i915)
> > @@ -8698,51 +8698,51 @@ static inline void intel_enable_llc_pstate(struct drm_i915_private *i915)
> >   	i915->gt_pm.llc_pstate.enabled = true;
> >   }
> > -static void intel_enable_rc6(struct drm_i915_private *dev_priv)
> > +static void intel_enable_rc6(struct drm_i915_private *i915)
> >   {
> > -	lockdep_assert_held(&dev_priv->gt_pm.rps.lock);
> > +	lockdep_assert_held(&i915->gt_pm.rps.lock);
> > -	if (dev_priv->gt_pm.rc6.enabled)
> > +	if (i915->gt_pm.rc6.enabled)
> >   		return;
> > -	if (IS_CHERRYVIEW(dev_priv))
> > -		cherryview_enable_rc6(dev_priv);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		valleyview_enable_rc6(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 11)
> > -		gen11_enable_rc6(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 9)
> > -		gen9_enable_rc6(dev_priv);
> > -	else if (IS_BROADWELL(dev_priv))
> > -		gen8_enable_rc6(dev_priv);
> > -	else if (INTEL_GEN(dev_priv) >= 6)
> > -		gen6_enable_rc6(dev_priv);
> > +	if (IS_CHERRYVIEW(i915))
> > +		cherryview_enable_rc6(i915);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		valleyview_enable_rc6(i915);
> > +	else if (INTEL_GEN(i915) >= 11)
> > +		gen11_enable_rc6(i915);
> > +	else if (INTEL_GEN(i915) >= 9)
> > +		gen9_enable_rc6(i915);
> > +	else if (IS_BROADWELL(i915))
> > +		gen8_enable_rc6(i915);
> > +	else if (INTEL_GEN(i915) >= 6)
> > +		gen6_enable_rc6(i915);
> > -	dev_priv->gt_pm.rc6.enabled = true;
> > +	i915->gt_pm.rc6.enabled = true;
> >   }
> > -static void intel_enable_rps(struct drm_i915_private *dev_priv)
> > +static void intel_enable_rps(struct drm_i915_private *i915)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	lockdep_assert_held(&rps->lock);
> >   	if (rps->enabled)
> >   		return;
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > -		cherryview_enable_rps(dev_priv);
> > -	} else if (IS_VALLEYVIEW(dev_priv)) {
> > -		valleyview_enable_rps(dev_priv);
> > -	} else if (INTEL_GEN(dev_priv) >= 9) {
> > -		gen9_enable_rps(dev_priv);
> > -	} else if (IS_BROADWELL(dev_priv)) {
> > -		gen8_enable_rps(dev_priv);
> > -	} else if (INTEL_GEN(dev_priv) >= 6) {
> > -		gen6_enable_rps(dev_priv);
> > -	} else if (IS_IRONLAKE_M(dev_priv)) {
> > -		ironlake_enable_drps(dev_priv);
> > -		intel_init_emon(dev_priv);
> > +	if (IS_CHERRYVIEW(i915)) {
> > +		cherryview_enable_rps(i915);
> > +	} else if (IS_VALLEYVIEW(i915)) {
> > +		valleyview_enable_rps(i915);
> > +	} else if (INTEL_GEN(i915) >= 9) {
> > +		gen9_enable_rps(i915);
> > +	} else if (IS_BROADWELL(i915)) {
> > +		gen8_enable_rps(i915);
> > +	} else if (INTEL_GEN(i915) >= 6) {
> > +		gen6_enable_rps(i915);
> > +	} else if (IS_IRONLAKE_M(i915)) {
> > +		ironlake_enable_drps(i915);
> > +		intel_init_emon(i915);
> >   	}
> >   	WARN_ON(rps->max_freq < rps->min_freq);
> > @@ -8754,25 +8754,25 @@ static void intel_enable_rps(struct drm_i915_private *dev_priv)
> >   	rps->enabled = true;
> >   }
> > -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv)
> > +void intel_enable_gt_powersave(struct drm_i915_private *i915)
> >   {
> >   	/* Powersaving is controlled by the host when inside a VM */
> > -	if (intel_vgpu_active(dev_priv))
> > +	if (intel_vgpu_active(i915))
> >   		return;
> > -	mutex_lock(&dev_priv->gt_pm.rps.lock);
> > +	mutex_lock(&i915->gt_pm.rps.lock);
> > -	if (HAS_RC6(dev_priv))
> > -		intel_enable_rc6(dev_priv);
> > -	if (HAS_RPS(dev_priv))
> > -		intel_enable_rps(dev_priv);
> > -	if (HAS_LLC(dev_priv))
> > -		intel_enable_llc_pstate(dev_priv);
> > +	if (HAS_RC6(i915))
> > +		intel_enable_rc6(i915);
> > +	if (HAS_RPS(i915))
> > +		intel_enable_rps(i915);
> > +	if (HAS_LLC(i915))
> > +		intel_enable_llc_pstate(i915);
> > -	mutex_unlock(&dev_priv->gt_pm.rps.lock);
> > +	mutex_unlock(&i915->gt_pm.rps.lock);
> >   }
> > -static void ibx_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void ibx_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * On Ibex Peak and Cougar Point, we need to disable clock
> > @@ -8782,11 +8782,11 @@ static void ibx_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(SOUTH_DSPCLK_GATE_D, PCH_DPLSUNIT_CLOCK_GATE_DISABLE);
> >   }
> > -static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv)
> > +static void g4x_disable_trickle_feed(struct drm_i915_private *i915)
> >   {
> >   	enum pipe pipe;
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		I915_WRITE(DSPCNTR(pipe),
> >   			   I915_READ(DSPCNTR(pipe)) |
> >   			   DISPPLANE_TRICKLE_FEED_DISABLE);
> > @@ -8796,7 +8796,7 @@ static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void ilk_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE;
> > @@ -8836,7 +8836,7 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	 * The bit 22 of 0x42004
> >   	 * The bit 7,8,9 of 0x42020.
> >   	 */
> > -	if (IS_IRONLAKE_M(dev_priv)) {
> > +	if (IS_IRONLAKE_M(i915)) {
> >   		/* WaFbcAsynchFlipDisableFbcQueue:ilk */
> >   		I915_WRITE(ILK_DISPLAY_CHICKEN1,
> >   			   I915_READ(ILK_DISPLAY_CHICKEN1) |
> > @@ -8862,12 +8862,12 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	/* WaDisable_RenderCache_OperationalFlush:ilk */
> >   	I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> > -	g4x_disable_trickle_feed(dev_priv);
> > +	g4x_disable_trickle_feed(i915);
> > -	ibx_init_clock_gating(dev_priv);
> > +	ibx_init_clock_gating(i915);
> >   }
> > -static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void cpt_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	int pipe;
> >   	u32 val;
> > @@ -8885,11 +8885,11 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	/* The below fixes the weird display corruption, a few pixels shifted
> >   	 * downward, on (only) LVDS of some HP laptops with IVY.
> >   	 */
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		val = I915_READ(TRANS_CHICKEN2(pipe));
> >   		val |= TRANS_CHICKEN2_TIMING_OVERRIDE;
> >   		val &= ~TRANS_CHICKEN2_FDI_POLARITY_REVERSED;
> > -		if (dev_priv->vbt.fdi_rx_polarity_inverted)
> > +		if (i915->vbt.fdi_rx_polarity_inverted)
> >   			val |= TRANS_CHICKEN2_FDI_POLARITY_REVERSED;
> >   		val &= ~TRANS_CHICKEN2_FRAME_START_DELAY_MASK;
> >   		val &= ~TRANS_CHICKEN2_DISABLE_DEEP_COLOR_COUNTER;
> > @@ -8897,13 +8897,13 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		I915_WRITE(TRANS_CHICKEN2(pipe), val);
> >   	}
> >   	/* WADP0ClockGatingDisable */
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		I915_WRITE(TRANS_CHICKEN1(pipe),
> >   			   TRANS_CHICKEN1_DP0UNIT_GC_DISABLE);
> >   	}
> >   }
> > -static void gen6_check_mch_setup(struct drm_i915_private *dev_priv)
> > +static void gen6_check_mch_setup(struct drm_i915_private *i915)
> >   {
> >   	u32 tmp;
> > @@ -8913,7 +8913,7 @@ static void gen6_check_mch_setup(struct drm_i915_private *dev_priv)
> >   			      tmp);
> >   }
> > -static void gen6_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void gen6_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE;
> > @@ -9000,14 +9000,14 @@ static void gen6_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   ILK_DPARBUNIT_CLOCK_GATE_ENABLE  |
> >   		   ILK_DPFDUNIT_CLOCK_GATE_ENABLE);
> > -	g4x_disable_trickle_feed(dev_priv);
> > +	g4x_disable_trickle_feed(i915);
> > -	cpt_init_clock_gating(dev_priv);
> > +	cpt_init_clock_gating(i915);
> > -	gen6_check_mch_setup(dev_priv);
> > +	gen6_check_mch_setup(i915);
> >   }
> > -static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv)
> > +static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *i915)
> >   {
> >   	u32 reg = I915_READ(GEN7_FF_THREAD_MODE);
> > @@ -9025,13 +9025,13 @@ static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(GEN7_FF_THREAD_MODE, reg);
> >   }
> > -static void lpt_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void lpt_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * TODO: this bit should only be enabled when really needed, then
> >   	 * disabled when not needed anymore in order to save power.
> >   	 */
> > -	if (HAS_PCH_LPT_LP(dev_priv))
> > +	if (HAS_PCH_LPT_LP(i915))
> >   		I915_WRITE(SOUTH_DSPCLK_GATE_D,
> >   			   I915_READ(SOUTH_DSPCLK_GATE_D) |
> >   			   PCH_LP_PARTITION_LEVEL_DISABLE);
> > @@ -9042,9 +9042,9 @@ static void lpt_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   TRANS_CHICKEN1_DP0UNIT_GC_DISABLE);
> >   }
> > -static void lpt_suspend_hw(struct drm_i915_private *dev_priv)
> > +static void lpt_suspend_hw(struct drm_i915_private *i915)
> >   {
> > -	if (HAS_PCH_LPT_LP(dev_priv)) {
> > +	if (HAS_PCH_LPT_LP(i915)) {
> >   		u32 val = I915_READ(SOUTH_DSPCLK_GATE_D);
> >   		val &= ~PCH_LP_PARTITION_LEVEL_DISABLE;
> > @@ -9052,7 +9052,7 @@ static void lpt_suspend_hw(struct drm_i915_private *dev_priv)
> >   	}
> >   }
> > -static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv,
> > +static void gen8_set_l3sqc_credits(struct drm_i915_private *i915,
> >   				   int general_prio_credits,
> >   				   int high_prio_credits)
> >   {
> > @@ -9078,7 +9078,7 @@ static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(GEN7_MISCCPCTL, misccpctl);
> >   }
> > -static void icl_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void icl_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/* This is not an Wa. Enable to reduce Sampler power */
> >   	I915_WRITE(GEN10_DFR_RATIO_EN_AND_CHICKEN,
> > @@ -9089,9 +9089,9 @@ static void icl_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   _MASKED_BIT_ENABLE(GEN11_ENABLE_32_PLANE_MODE));
> >   }
> > -static void cnp_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void cnp_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	if (!HAS_PCH_CNP(dev_priv))
> > +	if (!HAS_PCH_CNP(i915))
> >   		return;
> >   	/* Display WA #1181 WaSouthDisplayDisablePWMCGEGating: cnp */
> > @@ -9099,10 +9099,10 @@ static void cnp_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   CNP_PWM_CGE_GATING_DISABLE);
> >   }
> > -static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void cnl_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	cnp_init_clock_gating(dev_priv);
> > +	cnp_init_clock_gating(i915);
> >   	/* This is not an Wa. Enable for better image quality */
> >   	I915_WRITE(_3D_CHICKEN3,
> > @@ -9120,7 +9120,7 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	/* ReadHitWriteOnlyDisable:cnl */
> >   	val |= RCCUNIT_CLKGATE_DIS;
> >   	/* WaSarbUnitClockGatingDisable:cnl (pre-prod) */
> > -	if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_B0))
> > +	if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_B0))
> >   		val |= SARBUNIT_CLKGATE_DIS;
> >   	I915_WRITE(SLICE_UNIT_LEVEL_CLKGATE, val);
> > @@ -9136,27 +9136,27 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(UNSLICE_UNIT_LEVEL_CLKGATE, val);
> >   }
> > -static void cfl_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void cfl_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	cnp_init_clock_gating(dev_priv);
> > -	gen9_init_clock_gating(dev_priv);
> > +	cnp_init_clock_gating(i915);
> > +	gen9_init_clock_gating(i915);
> >   	/* WaFbcNukeOnHostModify:cfl */
> >   	I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) |
> >   		   ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> >   }
> > -static void kbl_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void kbl_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	gen9_init_clock_gating(dev_priv);
> > +	gen9_init_clock_gating(i915);
> >   	/* WaDisableSDEUnitClockGating:kbl */
> > -	if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0))
> > +	if (IS_KBL_REVID(i915, 0, KBL_REVID_B0))
> >   		I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) |
> >   			   GEN8_SDEUNIT_CLOCK_GATE_DISABLE);
> >   	/* WaDisableGamClockGating:kbl */
> > -	if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0))
> > +	if (IS_KBL_REVID(i915, 0, KBL_REVID_B0))
> >   		I915_WRITE(GEN6_UCGCTL1, I915_READ(GEN6_UCGCTL1) |
> >   			   GEN6_GAMUNIT_CLOCK_GATE_DISABLE);
> > @@ -9165,9 +9165,9 @@ static void kbl_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> >   }
> > -static void skl_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void skl_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	gen9_init_clock_gating(dev_priv);
> > +	gen9_init_clock_gating(i915);
> >   	/* WAC6entrylatency:skl */
> >   	I915_WRITE(FBC_LLC_READ_CTRL, I915_READ(FBC_LLC_READ_CTRL) |
> > @@ -9178,10 +9178,10 @@ static void skl_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   ILK_DPFC_NUKE_ON_ANY_MODIFICATION);
> >   }
> > -static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void bdw_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/* The GTT cache must be disabled if the system is using 2M pages. */
> > -	bool can_use_gtt_cache = !HAS_PAGE_SIZES(dev_priv,
> > +	bool can_use_gtt_cache = !HAS_PAGE_SIZES(i915,
> >   						 I915_GTT_PAGE_SIZE_2M);
> >   	enum pipe pipe;
> > @@ -9193,7 +9193,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   I915_READ(CHICKEN_PAR1_1) | DPA_MASK_VBLANK_SRD);
> >   	/* WaPsrDPRSUnmaskVBlankInSRD:bdw */
> > -	for_each_pipe(dev_priv, pipe) {
> > +	for_each_pipe(i915, pipe) {
> >   		I915_WRITE(CHICKEN_PIPESL_1(pipe),
> >   			   I915_READ(CHICKEN_PIPESL_1(pipe)) |
> >   			   BDW_DPRS_MASK_VBLANK_SRD);
> > @@ -9213,7 +9213,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   GEN8_SDEUNIT_CLOCK_GATE_DISABLE);
> >   	/* WaProgramL3SqcReg1Default:bdw */
> > -	gen8_set_l3sqc_credits(dev_priv, 30, 2);
> > +	gen8_set_l3sqc_credits(i915, 30, 2);
> >   	/* WaGttCachingOffByDefault:bdw */
> >   	I915_WRITE(HSW_GTT_CACHE_EN, can_use_gtt_cache ? GTT_CACHE_EN_ALL : 0);
> > @@ -9222,7 +9222,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(CHICKEN_PAR2_1, I915_READ(CHICKEN_PAR2_1)
> >   		   | KVM_CONFIG_CHANGE_NOTIFICATION_SELECT);
> > -	lpt_init_clock_gating(dev_priv);
> > +	lpt_init_clock_gating(i915);
> >   	/* WaDisableDopClockGating:bdw
> >   	 *
> > @@ -9233,7 +9233,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   I915_READ(GEN6_UCGCTL1) | GEN6_EU_TCUNIT_CLOCK_GATE_DISABLE);
> >   }
> > -static void hsw_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void hsw_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/* L3 caching of data atomics doesn't work -- disable it. */
> >   	I915_WRITE(HSW_SCRATCH1, HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE);
> > @@ -9278,10 +9278,10 @@ static void hsw_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	/* WaSwitchSolVfFArbitrationPriority:hsw */
> >   	I915_WRITE(GAM_ECOCHK, I915_READ(GAM_ECOCHK) | HSW_ECOCHK_ARB_PRIO_SOL);
> > -	lpt_init_clock_gating(dev_priv);
> > +	lpt_init_clock_gating(i915);
> >   }
> > -static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void ivb_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 snpcr;
> > @@ -9297,7 +9297,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   CHICKEN3_DGMG_DONE_FIX_DISABLE);
> >   	/* WaDisablePSDDualDispatchEnable:ivb */
> > -	if (IS_IVB_GT1(dev_priv))
> > +	if (IS_IVB_GT1(i915))
> >   		I915_WRITE(GEN7_HALF_SLICE_CHICKEN1,
> >   			   _MASKED_BIT_ENABLE(GEN7_PSD_SINGLE_PORT_DISPATCH_ENABLE));
> > @@ -9313,7 +9313,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> >   			GEN7_WA_FOR_GEN7_L3_CONTROL);
> >   	I915_WRITE(GEN7_L3_CHICKEN_MODE_REGISTER,
> >   		   GEN7_WA_L3_CHICKEN_MODE);
> > -	if (IS_IVB_GT1(dev_priv))
> > +	if (IS_IVB_GT1(i915))
> >   		I915_WRITE(GEN7_ROW_CHICKEN2,
> >   			   _MASKED_BIT_ENABLE(DOP_CLOCK_GATING_DISABLE));
> >   	else {
> > @@ -9340,9 +9340,9 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> >   			I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) |
> >   			GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB);
> > -	g4x_disable_trickle_feed(dev_priv);
> > +	g4x_disable_trickle_feed(i915);
> > -	gen7_setup_fixed_func_scheduler(dev_priv);
> > +	gen7_setup_fixed_func_scheduler(i915);
> >   	if (0) { /* causes HiZ corruption on ivb:gt1 */
> >   		/* enable HiZ Raw Stall Optimization */
> > @@ -9370,13 +9370,13 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	snpcr |= GEN6_MBC_SNPCR_MED;
> >   	I915_WRITE(GEN6_MBCUNIT_SNPCR, snpcr);
> > -	if (!HAS_PCH_NOP(dev_priv))
> > -		cpt_init_clock_gating(dev_priv);
> > +	if (!HAS_PCH_NOP(i915))
> > +		cpt_init_clock_gating(i915);
> > -	gen6_check_mch_setup(dev_priv);
> > +	gen6_check_mch_setup(i915);
> >   }
> > -static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void vlv_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/* WaDisableEarlyCull:vlv */
> >   	I915_WRITE(_3D_CHICKEN3,
> > @@ -9409,7 +9409,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) |
> >   		   GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB);
> > -	gen7_setup_fixed_func_scheduler(dev_priv);
> > +	gen7_setup_fixed_func_scheduler(i915);
> >   	/*
> >   	 * According to the spec, bit 13 (RCZUNIT) must be set on IVB.
> > @@ -9456,7 +9456,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(VLV_GUNIT_CLOCK_GATE, GCFG_DIS);
> >   }
> > -static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void chv_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	/* WaVSRefCountFullforceMissDisable:chv */
> >   	/* WaDSRefCountFullforceMissDisable:chv */
> > @@ -9481,7 +9481,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	 * See gfxspecs/Related Documents/Performance Guide/
> >   	 * LSQC Setting Recommendations.
> >   	 */
> > -	gen8_set_l3sqc_credits(dev_priv, 38, 2);
> > +	gen8_set_l3sqc_credits(i915, 38, 2);
> >   	/*
> >   	 * GTT cache may not work with big pages, so if those
> > @@ -9490,7 +9490,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(HSW_GTT_CACHE_EN, GTT_CACHE_EN_ALL);
> >   }
> > -static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void g4x_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 dspclk_gate;
> > @@ -9502,7 +9502,7 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	dspclk_gate = VRHUNIT_CLOCK_GATE_DISABLE |
> >   		OVRUNIT_CLOCK_GATE_DISABLE |
> >   		OVCUNIT_CLOCK_GATE_DISABLE;
> > -	if (IS_GM45(dev_priv))
> > +	if (IS_GM45(i915))
> >   		dspclk_gate |= DSSUNIT_CLOCK_GATE_DISABLE;
> >   	I915_WRITE(DSPCLK_GATE_D, dspclk_gate);
> > @@ -9513,12 +9513,12 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	/* WaDisable_RenderCache_OperationalFlush:g4x */
> >   	I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> > -	g4x_disable_trickle_feed(dev_priv);
> > +	g4x_disable_trickle_feed(i915);
> >   }
> > -static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void i965gm_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	intel_uncore_write(uncore, RENCLK_GATE_D1, I965_RCC_CLOCK_GATE_DISABLE);
> >   	intel_uncore_write(uncore, RENCLK_GATE_D2, 0);
> > @@ -9535,7 +9535,7 @@ static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv)
> >   			   _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> >   }
> > -static void i965g_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void i965g_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(RENCLK_GATE_D1, I965_RCZ_CLOCK_GATE_DISABLE |
> >   		   I965_RCC_CLOCK_GATE_DISABLE |
> > @@ -9550,7 +9550,7 @@ static void i965g_init_clock_gating(struct drm_i915_private *dev_priv)
> >   	I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE));
> >   }
> > -static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void gen3_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	u32 dstate = I915_READ(D_STATE);
> > @@ -9558,7 +9558,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		DSTATE_DOT_CLOCK_GATING;
> >   	I915_WRITE(D_STATE, dstate);
> > -	if (IS_PINEVIEW(dev_priv))
> > +	if (IS_PINEVIEW(i915))
> >   		I915_WRITE(ECOSKPD, _MASKED_BIT_ENABLE(ECO_GATING_CX_ONLY));
> >   	/* IIR "flip pending" means done if this bit is set */
> > @@ -9574,7 +9574,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   _MASKED_BIT_ENABLE(MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE));
> >   }
> > -static void i85x_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void i85x_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(RENCLK_GATE_D1, SV_CLOCK_GATE_DISABLE);
> > @@ -9586,168 +9586,168 @@ static void i85x_init_clock_gating(struct drm_i915_private *dev_priv)
> >   		   _MASKED_BIT_ENABLE(MEM_DISPLAY_TRICKLE_FEED_DISABLE));
> >   }
> > -static void i830_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void i830_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	I915_WRITE(MEM_MODE,
> >   		   _MASKED_BIT_ENABLE(MEM_DISPLAY_A_TRICKLE_FEED_DISABLE) |
> >   		   _MASKED_BIT_ENABLE(MEM_DISPLAY_B_TRICKLE_FEED_DISABLE));
> >   }
> > -void intel_init_clock_gating(struct drm_i915_private *dev_priv)
> > +void intel_init_clock_gating(struct drm_i915_private *i915)
> >   {
> > -	dev_priv->display.init_clock_gating(dev_priv);
> > +	i915->display.init_clock_gating(i915);
> >   }
> > -void intel_suspend_hw(struct drm_i915_private *dev_priv)
> > +void intel_suspend_hw(struct drm_i915_private *i915)
> >   {
> > -	if (HAS_PCH_LPT(dev_priv))
> > -		lpt_suspend_hw(dev_priv);
> > +	if (HAS_PCH_LPT(i915))
> > +		lpt_suspend_hw(i915);
> >   }
> > -static void nop_init_clock_gating(struct drm_i915_private *dev_priv)
> > +static void nop_init_clock_gating(struct drm_i915_private *i915)
> >   {
> >   	DRM_DEBUG_KMS("No clock gating settings or workarounds applied.\n");
> >   }
> >   /**
> >    * intel_init_clock_gating_hooks - setup the clock gating hooks
> > - * @dev_priv: device private
> > + * @i915: device private
> >    *
> >    * Setup the hooks that configure which clocks of a given platform can be
> >    * gated and also apply various GT and display specific workarounds for these
> >    * platforms. Note that some GT specific workarounds are applied separately
> >    * when GPU contexts or batchbuffers start their execution.
> >    */
> > -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv)
> > -{
> > -	if (IS_GEN(dev_priv, 11))
> > -		dev_priv->display.init_clock_gating = icl_init_clock_gating;
> > -	else if (IS_CANNONLAKE(dev_priv))
> > -		dev_priv->display.init_clock_gating = cnl_init_clock_gating;
> > -	else if (IS_COFFEELAKE(dev_priv))
> > -		dev_priv->display.init_clock_gating = cfl_init_clock_gating;
> > -	else if (IS_SKYLAKE(dev_priv))
> > -		dev_priv->display.init_clock_gating = skl_init_clock_gating;
> > -	else if (IS_KABYLAKE(dev_priv))
> > -		dev_priv->display.init_clock_gating = kbl_init_clock_gating;
> > -	else if (IS_BROXTON(dev_priv))
> > -		dev_priv->display.init_clock_gating = bxt_init_clock_gating;
> > -	else if (IS_GEMINILAKE(dev_priv))
> > -		dev_priv->display.init_clock_gating = glk_init_clock_gating;
> > -	else if (IS_BROADWELL(dev_priv))
> > -		dev_priv->display.init_clock_gating = bdw_init_clock_gating;
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		dev_priv->display.init_clock_gating = chv_init_clock_gating;
> > -	else if (IS_HASWELL(dev_priv))
> > -		dev_priv->display.init_clock_gating = hsw_init_clock_gating;
> > -	else if (IS_IVYBRIDGE(dev_priv))
> > -		dev_priv->display.init_clock_gating = ivb_init_clock_gating;
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		dev_priv->display.init_clock_gating = vlv_init_clock_gating;
> > -	else if (IS_GEN(dev_priv, 6))
> > -		dev_priv->display.init_clock_gating = gen6_init_clock_gating;
> > -	else if (IS_GEN(dev_priv, 5))
> > -		dev_priv->display.init_clock_gating = ilk_init_clock_gating;
> > -	else if (IS_G4X(dev_priv))
> > -		dev_priv->display.init_clock_gating = g4x_init_clock_gating;
> > -	else if (IS_I965GM(dev_priv))
> > -		dev_priv->display.init_clock_gating = i965gm_init_clock_gating;
> > -	else if (IS_I965G(dev_priv))
> > -		dev_priv->display.init_clock_gating = i965g_init_clock_gating;
> > -	else if (IS_GEN(dev_priv, 3))
> > -		dev_priv->display.init_clock_gating = gen3_init_clock_gating;
> > -	else if (IS_I85X(dev_priv) || IS_I865G(dev_priv))
> > -		dev_priv->display.init_clock_gating = i85x_init_clock_gating;
> > -	else if (IS_GEN(dev_priv, 2))
> > -		dev_priv->display.init_clock_gating = i830_init_clock_gating;
> > +void intel_init_clock_gating_hooks(struct drm_i915_private *i915)
> > +{
> > +	if (IS_GEN(i915, 11))
> > +		i915->display.init_clock_gating = icl_init_clock_gating;
> > +	else if (IS_CANNONLAKE(i915))
> > +		i915->display.init_clock_gating = cnl_init_clock_gating;
> > +	else if (IS_COFFEELAKE(i915))
> > +		i915->display.init_clock_gating = cfl_init_clock_gating;
> > +	else if (IS_SKYLAKE(i915))
> > +		i915->display.init_clock_gating = skl_init_clock_gating;
> > +	else if (IS_KABYLAKE(i915))
> > +		i915->display.init_clock_gating = kbl_init_clock_gating;
> > +	else if (IS_BROXTON(i915))
> > +		i915->display.init_clock_gating = bxt_init_clock_gating;
> > +	else if (IS_GEMINILAKE(i915))
> > +		i915->display.init_clock_gating = glk_init_clock_gating;
> > +	else if (IS_BROADWELL(i915))
> > +		i915->display.init_clock_gating = bdw_init_clock_gating;
> > +	else if (IS_CHERRYVIEW(i915))
> > +		i915->display.init_clock_gating = chv_init_clock_gating;
> > +	else if (IS_HASWELL(i915))
> > +		i915->display.init_clock_gating = hsw_init_clock_gating;
> > +	else if (IS_IVYBRIDGE(i915))
> > +		i915->display.init_clock_gating = ivb_init_clock_gating;
> > +	else if (IS_VALLEYVIEW(i915))
> > +		i915->display.init_clock_gating = vlv_init_clock_gating;
> > +	else if (IS_GEN(i915, 6))
> > +		i915->display.init_clock_gating = gen6_init_clock_gating;
> > +	else if (IS_GEN(i915, 5))
> > +		i915->display.init_clock_gating = ilk_init_clock_gating;
> > +	else if (IS_G4X(i915))
> > +		i915->display.init_clock_gating = g4x_init_clock_gating;
> > +	else if (IS_I965GM(i915))
> > +		i915->display.init_clock_gating = i965gm_init_clock_gating;
> > +	else if (IS_I965G(i915))
> > +		i915->display.init_clock_gating = i965g_init_clock_gating;
> > +	else if (IS_GEN(i915, 3))
> > +		i915->display.init_clock_gating = gen3_init_clock_gating;
> > +	else if (IS_I85X(i915) || IS_I865G(i915))
> > +		i915->display.init_clock_gating = i85x_init_clock_gating;
> > +	else if (IS_GEN(i915, 2))
> > +		i915->display.init_clock_gating = i830_init_clock_gating;
> >   	else {
> > -		MISSING_CASE(INTEL_DEVID(dev_priv));
> > -		dev_priv->display.init_clock_gating = nop_init_clock_gating;
> > +		MISSING_CASE(INTEL_DEVID(i915));
> > +		i915->display.init_clock_gating = nop_init_clock_gating;
> >   	}
> >   }
> >   /* Set up chip specific power management-related functions */
> > -void intel_init_pm(struct drm_i915_private *dev_priv)
> > +void intel_init_pm(struct drm_i915_private *i915)
> >   {
> >   	/* For cxsr */
> > -	if (IS_PINEVIEW(dev_priv))
> > -		i915_pineview_get_mem_freq(dev_priv);
> > -	else if (IS_GEN(dev_priv, 5))
> > -		i915_ironlake_get_mem_freq(dev_priv);
> > +	if (IS_PINEVIEW(i915))
> > +		i915_pineview_get_mem_freq(i915);
> > +	else if (IS_GEN(i915, 5))
> > +		i915_ironlake_get_mem_freq(i915);
> >   	/* For FIFO watermark updates */
> > -	if (INTEL_GEN(dev_priv) >= 9) {
> > -		skl_setup_wm_latency(dev_priv);
> > -		dev_priv->display.initial_watermarks = skl_initial_wm;
> > -		dev_priv->display.atomic_update_watermarks = skl_atomic_update_crtc_wm;
> > -		dev_priv->display.compute_global_watermarks = skl_compute_wm;
> > -	} else if (HAS_PCH_SPLIT(dev_priv)) {
> > -		ilk_setup_wm_latency(dev_priv);
> > -
> > -		if ((IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[1] &&
> > -		     dev_priv->wm.spr_latency[1] && dev_priv->wm.cur_latency[1]) ||
> > -		    (!IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[0] &&
> > -		     dev_priv->wm.spr_latency[0] && dev_priv->wm.cur_latency[0])) {
> > -			dev_priv->display.compute_pipe_wm = ilk_compute_pipe_wm;
> > -			dev_priv->display.compute_intermediate_wm =
> > +	if (INTEL_GEN(i915) >= 9) {
> > +		skl_setup_wm_latency(i915);
> > +		i915->display.initial_watermarks = skl_initial_wm;
> > +		i915->display.atomic_update_watermarks = skl_atomic_update_crtc_wm;
> > +		i915->display.compute_global_watermarks = skl_compute_wm;
> > +	} else if (HAS_PCH_SPLIT(i915)) {
> > +		ilk_setup_wm_latency(i915);
> > +
> > +		if ((IS_GEN(i915, 5) && i915->wm.pri_latency[1] &&
> > +		     i915->wm.spr_latency[1] && i915->wm.cur_latency[1]) ||
> > +		    (!IS_GEN(i915, 5) && i915->wm.pri_latency[0] &&
> > +		     i915->wm.spr_latency[0] && i915->wm.cur_latency[0])) {
> > +			i915->display.compute_pipe_wm = ilk_compute_pipe_wm;
> > +			i915->display.compute_intermediate_wm =
> >   				ilk_compute_intermediate_wm;
> > -			dev_priv->display.initial_watermarks =
> > +			i915->display.initial_watermarks =
> >   				ilk_initial_watermarks;
> > -			dev_priv->display.optimize_watermarks =
> > +			i915->display.optimize_watermarks =
> >   				ilk_optimize_watermarks;
> >   		} else {
> >   			DRM_DEBUG_KMS("Failed to read display plane latency. "
> >   				      "Disable CxSR\n");
> >   		}
> > -	} else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > -		vlv_setup_wm_latency(dev_priv);
> > -		dev_priv->display.compute_pipe_wm = vlv_compute_pipe_wm;
> > -		dev_priv->display.compute_intermediate_wm = vlv_compute_intermediate_wm;
> > -		dev_priv->display.initial_watermarks = vlv_initial_watermarks;
> > -		dev_priv->display.optimize_watermarks = vlv_optimize_watermarks;
> > -		dev_priv->display.atomic_update_watermarks = vlv_atomic_update_fifo;
> > -	} else if (IS_G4X(dev_priv)) {
> > -		g4x_setup_wm_latency(dev_priv);
> > -		dev_priv->display.compute_pipe_wm = g4x_compute_pipe_wm;
> > -		dev_priv->display.compute_intermediate_wm = g4x_compute_intermediate_wm;
> > -		dev_priv->display.initial_watermarks = g4x_initial_watermarks;
> > -		dev_priv->display.optimize_watermarks = g4x_optimize_watermarks;
> > -	} else if (IS_PINEVIEW(dev_priv)) {
> > -		if (!intel_get_cxsr_latency(!IS_MOBILE(dev_priv),
> > -					    dev_priv->is_ddr3,
> > -					    dev_priv->fsb_freq,
> > -					    dev_priv->mem_freq)) {
> > +	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> > +		vlv_setup_wm_latency(i915);
> > +		i915->display.compute_pipe_wm = vlv_compute_pipe_wm;
> > +		i915->display.compute_intermediate_wm = vlv_compute_intermediate_wm;
> > +		i915->display.initial_watermarks = vlv_initial_watermarks;
> > +		i915->display.optimize_watermarks = vlv_optimize_watermarks;
> > +		i915->display.atomic_update_watermarks = vlv_atomic_update_fifo;
> > +	} else if (IS_G4X(i915)) {
> > +		g4x_setup_wm_latency(i915);
> > +		i915->display.compute_pipe_wm = g4x_compute_pipe_wm;
> > +		i915->display.compute_intermediate_wm = g4x_compute_intermediate_wm;
> > +		i915->display.initial_watermarks = g4x_initial_watermarks;
> > +		i915->display.optimize_watermarks = g4x_optimize_watermarks;
> > +	} else if (IS_PINEVIEW(i915)) {
> > +		if (!intel_get_cxsr_latency(!IS_MOBILE(i915),
> > +					    i915->is_ddr3,
> > +					    i915->fsb_freq,
> > +					    i915->mem_freq)) {
> >   			DRM_INFO("failed to find known CxSR latency "
> >   				 "(found ddr%s fsb freq %d, mem freq %d), "
> >   				 "disabling CxSR\n",
> > -				 (dev_priv->is_ddr3 == 1) ? "3" : "2",
> > -				 dev_priv->fsb_freq, dev_priv->mem_freq);
> > +				 (i915->is_ddr3 == 1) ? "3" : "2",
> > +				 i915->fsb_freq, i915->mem_freq);
> >   			/* Disable CxSR and never update its watermark again */
> > -			intel_set_memory_cxsr(dev_priv, false);
> > -			dev_priv->display.update_wm = NULL;
> > +			intel_set_memory_cxsr(i915, false);
> > +			i915->display.update_wm = NULL;
> >   		} else
> > -			dev_priv->display.update_wm = pineview_update_wm;
> > -	} else if (IS_GEN(dev_priv, 4)) {
> > -		dev_priv->display.update_wm = i965_update_wm;
> > -	} else if (IS_GEN(dev_priv, 3)) {
> > -		dev_priv->display.update_wm = i9xx_update_wm;
> > -		dev_priv->display.get_fifo_size = i9xx_get_fifo_size;
> > -	} else if (IS_GEN(dev_priv, 2)) {
> > -		if (INTEL_INFO(dev_priv)->num_pipes == 1) {
> > -			dev_priv->display.update_wm = i845_update_wm;
> > -			dev_priv->display.get_fifo_size = i845_get_fifo_size;
> > +			i915->display.update_wm = pineview_update_wm;
> > +	} else if (IS_GEN(i915, 4)) {
> > +		i915->display.update_wm = i965_update_wm;
> > +	} else if (IS_GEN(i915, 3)) {
> > +		i915->display.update_wm = i9xx_update_wm;
> > +		i915->display.get_fifo_size = i9xx_get_fifo_size;
> > +	} else if (IS_GEN(i915, 2)) {
> > +		if (INTEL_INFO(i915)->num_pipes == 1) {
> > +			i915->display.update_wm = i845_update_wm;
> > +			i915->display.get_fifo_size = i845_get_fifo_size;
> >   		} else {
> > -			dev_priv->display.update_wm = i9xx_update_wm;
> > -			dev_priv->display.get_fifo_size = i830_get_fifo_size;
> > +			i915->display.update_wm = i9xx_update_wm;
> > +			i915->display.get_fifo_size = i830_get_fifo_size;
> >   		}
> >   	} else {
> >   		DRM_ERROR("unexpected fall-through in intel_init_pm\n");
> >   	}
> >   }
> > -static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val)
> > +static int byt_gpu_freq(struct drm_i915_private *i915, int val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/*
> >   	 * N = val - 0xb7
> > @@ -9756,16 +9756,16 @@ static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val)
> >   	return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * (val - 0xb7), 1000);
> >   }
> > -static int byt_freq_opcode(struct drm_i915_private *dev_priv, int val)
> > +static int byt_freq_opcode(struct drm_i915_private *i915, int val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	return DIV_ROUND_CLOSEST(1000 * val, rps->gpll_ref_freq) + 0xb7;
> >   }
> > -static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val)
> > +static int chv_gpu_freq(struct drm_i915_private *i915, int val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/*
> >   	 * N = val / 2
> > @@ -9774,52 +9774,52 @@ static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val)
> >   	return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * val, 2 * 2 * 1000);
> >   }
> > -static int chv_freq_opcode(struct drm_i915_private *dev_priv, int val)
> > +static int chv_freq_opcode(struct drm_i915_private *i915, int val)
> >   {
> > -	struct intel_rps *rps = &dev_priv->gt_pm.rps;
> > +	struct intel_rps *rps = &i915->gt_pm.rps;
> >   	/* CHV needs even values */
> >   	return DIV_ROUND_CLOSEST(2 * 1000 * val, rps->gpll_ref_freq) * 2;
> >   }
> > -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val)
> > +int intel_gpu_freq(struct drm_i915_private *i915, int val)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return DIV_ROUND_CLOSEST(val * GT_FREQUENCY_MULTIPLIER,
> >   					 GEN9_FREQ_SCALER);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		return chv_gpu_freq(dev_priv, val);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		return byt_gpu_freq(dev_priv, val);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		return chv_gpu_freq(i915, val);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		return byt_gpu_freq(i915, val);
> >   	else
> >   		return val * GT_FREQUENCY_MULTIPLIER;
> >   }
> > -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val)
> > +int intel_freq_opcode(struct drm_i915_private *i915, int val)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		return DIV_ROUND_CLOSEST(val * GEN9_FREQ_SCALER,
> >   					 GT_FREQUENCY_MULTIPLIER);
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > -		return chv_freq_opcode(dev_priv, val);
> > -	else if (IS_VALLEYVIEW(dev_priv))
> > -		return byt_freq_opcode(dev_priv, val);
> > +	else if (IS_CHERRYVIEW(i915))
> > +		return chv_freq_opcode(i915, val);
> > +	else if (IS_VALLEYVIEW(i915))
> > +		return byt_freq_opcode(i915, val);
> >   	else
> >   		return DIV_ROUND_CLOSEST(val, GT_FREQUENCY_MULTIPLIER);
> >   }
> > -void intel_pm_setup(struct drm_i915_private *dev_priv)
> > +void intel_pm_setup(struct drm_i915_private *i915)
> >   {
> > -	mutex_init(&dev_priv->gt_pm.rps.lock);
> > -	mutex_init(&dev_priv->gt_pm.rps.power.mutex);
> > +	mutex_init(&i915->gt_pm.rps.lock);
> > +	mutex_init(&i915->gt_pm.rps.power.mutex);
> > -	atomic_set(&dev_priv->gt_pm.rps.num_waiters, 0);
> > +	atomic_set(&i915->gt_pm.rps.num_waiters, 0);
> > -	dev_priv->runtime_pm.suspended = false;
> > -	atomic_set(&dev_priv->runtime_pm.wakeref_count, 0);
> > +	i915->runtime_pm.suspended = false;
> > +	atomic_set(&i915->runtime_pm.wakeref_count, 0);
> >   }
> > -static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> > +static u64 vlv_residency_raw(struct drm_i915_private *i915,
> >   			     const i915_reg_t reg)
> >   {
> >   	u32 lower, upper, tmp;
> > @@ -9829,7 +9829,7 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> >   	 * The register accessed do not need forcewake. We borrow
> >   	 * uncore lock to prevent concurrent access to range reg.
> >   	 */
> > -	lockdep_assert_held(&dev_priv->uncore.lock);
> > +	lockdep_assert_held(&i915->uncore.lock);
> >   	/*
> >   	 * vlv and chv residency counters are 40 bits in width.
> > @@ -9865,17 +9865,17 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv,
> >   	return lower | (u64)upper << 8;
> >   }
> > -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> > +u64 intel_rc6_residency_ns(struct drm_i915_private *i915,
> >   			   const i915_reg_t reg)
> >   {
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	u64 time_hw, prev_hw, overflow_hw;
> >   	unsigned int fw_domains;
> >   	unsigned long flags;
> >   	unsigned int i;
> >   	u32 mul, div;
> > -	if (!HAS_RC6(dev_priv))
> > +	if (!HAS_RC6(i915))
> >   		return 0;
> >   	/*
> > @@ -9887,7 +9887,7 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> >   	 */
> >   	i = (i915_mmio_reg_offset(reg) -
> >   	     i915_mmio_reg_offset(GEN6_GT_GFX_RC6_LOCKED)) / sizeof(u32);
> > -	if (WARN_ON_ONCE(i >= ARRAY_SIZE(dev_priv->gt_pm.rc6.cur_residency)))
> > +	if (WARN_ON_ONCE(i >= ARRAY_SIZE(i915->gt_pm.rc6.cur_residency)))
> >   		return 0;
> >   	fw_domains = intel_uncore_forcewake_for_reg(uncore, reg, FW_REG_READ);
> > @@ -9896,14 +9896,14 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> >   	intel_uncore_forcewake_get__locked(uncore, fw_domains);
> >   	/* On VLV and CHV, residency time is in CZ units rather than 1.28us */
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		mul = 1000000;
> > -		div = dev_priv->czclk_freq;
> > +		div = i915->czclk_freq;
> >   		overflow_hw = BIT_ULL(40);
> > -		time_hw = vlv_residency_raw(dev_priv, reg);
> > +		time_hw = vlv_residency_raw(i915, reg);
> >   	} else {
> >   		/* 833.33ns units on Gen9LP, 1.28us elsewhere. */
> > -		if (IS_GEN9_LP(dev_priv)) {
> > +		if (IS_GEN9_LP(i915)) {
> >   			mul = 10000;
> >   			div = 12;
> >   		} else {
> > @@ -9921,8 +9921,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> >   	 * But relying on a sufficient frequency of queries otherwise counters
> >   	 * can still wrap.
> >   	 */
> > -	prev_hw = dev_priv->gt_pm.rc6.prev_hw_residency[i];
> > -	dev_priv->gt_pm.rc6.prev_hw_residency[i] = time_hw;
> > +	prev_hw = i915->gt_pm.rc6.prev_hw_residency[i];
> > +	i915->gt_pm.rc6.prev_hw_residency[i] = time_hw;
> >   	/* RC6 delta from last sample. */
> >   	if (time_hw >= prev_hw)
> > @@ -9931,8 +9931,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> >   		time_hw += overflow_hw - prev_hw;
> >   	/* Add delta to RC6 extended raw driver copy. */
> > -	time_hw += dev_priv->gt_pm.rc6.cur_residency[i];
> > -	dev_priv->gt_pm.rc6.cur_residency[i] = time_hw;
> > +	time_hw += i915->gt_pm.rc6.cur_residency[i];
> > +	i915->gt_pm.rc6.cur_residency[i] = time_hw;
> >   	intel_uncore_forcewake_put__locked(uncore, fw_domains);
> >   	spin_unlock_irqrestore(&uncore->lock, flags);
> > @@ -9940,19 +9940,19 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv,
> >   	return mul_u64_u32_div(time_hw, mul, div);
> >   }
> > -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv,
> > +u64 intel_rc6_residency_us(struct drm_i915_private *i915,
> >   			   i915_reg_t reg)
> >   {
> > -	return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(dev_priv, reg), 1000);
> > +	return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(i915, reg), 1000);
> >   }
> > -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat)
> > +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat)
> >   {
> >   	u32 cagf;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		cagf = (rpstat & GEN9_CAGF_MASK) >> GEN9_CAGF_SHIFT;
> > -	else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	else if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		cagf = (rpstat & HSW_CAGF_MASK) >> HSW_CAGF_SHIFT;
> >   	else
> >   		cagf = (rpstat & GEN6_CAGF_MASK) >> GEN6_CAGF_SHIFT;
> > diff --git a/drivers/gpu/drm/i915/intel_pm.h b/drivers/gpu/drm/i915/intel_pm.h
> > index 1b489fa399e1..e2c4edf263f7 100644
> > --- a/drivers/gpu/drm/i915/intel_pm.h
> > +++ b/drivers/gpu/drm/i915/intel_pm.h
> > @@ -22,39 +22,39 @@ struct skl_ddb_entry;
> >   struct skl_pipe_wm;
> >   struct skl_wm_level;
> > -void intel_init_clock_gating(struct drm_i915_private *dev_priv);
> > -void intel_suspend_hw(struct drm_i915_private *dev_priv);
> > -int ilk_wm_max_level(const struct drm_i915_private *dev_priv);
> > +void intel_init_clock_gating(struct drm_i915_private *i915);
> > +void intel_suspend_hw(struct drm_i915_private *i915);
> > +int ilk_wm_max_level(const struct drm_i915_private *i915);
> >   void intel_update_watermarks(struct intel_crtc *crtc);
> > -void intel_init_pm(struct drm_i915_private *dev_priv);
> > -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv);
> > -void intel_pm_setup(struct drm_i915_private *dev_priv);
> > -void intel_gpu_ips_init(struct drm_i915_private *dev_priv);
> > +void intel_init_pm(struct drm_i915_private *i915);
> > +void intel_init_clock_gating_hooks(struct drm_i915_private *i915);
> > +void intel_pm_setup(struct drm_i915_private *i915);
> > +void intel_gpu_ips_init(struct drm_i915_private *i915);
> >   void intel_gpu_ips_teardown(void);
> > -void intel_init_gt_powersave(struct drm_i915_private *dev_priv);
> > -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv);
> > -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv);
> > -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv);
> > -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv);
> > -void gen6_rps_busy(struct drm_i915_private *dev_priv);
> > -void gen6_rps_idle(struct drm_i915_private *dev_priv);
> > +void intel_init_gt_powersave(struct drm_i915_private *i915);
> > +void intel_cleanup_gt_powersave(struct drm_i915_private *i915);
> > +void intel_sanitize_gt_powersave(struct drm_i915_private *i915);
> > +void intel_enable_gt_powersave(struct drm_i915_private *i915);
> > +void intel_disable_gt_powersave(struct drm_i915_private *i915);
> > +void gen6_rps_busy(struct drm_i915_private *i915);
> > +void gen6_rps_idle(struct drm_i915_private *i915);
> >   void gen6_rps_boost(struct i915_request *rq);
> > -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv);
> > -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv);
> > -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv);
> > -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv);
> > +void g4x_wm_get_hw_state(struct drm_i915_private *i915);
> > +void vlv_wm_get_hw_state(struct drm_i915_private *i915);
> > +void ilk_wm_get_hw_state(struct drm_i915_private *i915);
> > +void skl_wm_get_hw_state(struct drm_i915_private *i915);
> >   void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc,
> >   			       struct skl_ddb_entry *ddb_y,
> >   			       struct skl_ddb_entry *ddb_uv);
> > -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv,
> > +void skl_ddb_get_hw_state(struct drm_i915_private *i915,
> >   			  struct skl_ddb_allocation *ddb /* out */);
> >   void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc,
> >   			      struct skl_pipe_wm *out);
> > -void g4x_wm_sanitize(struct drm_i915_private *dev_priv);
> > -void vlv_wm_sanitize(struct drm_i915_private *dev_priv);
> > +void g4x_wm_sanitize(struct drm_i915_private *i915);
> > +void vlv_wm_sanitize(struct drm_i915_private *i915);
> >   bool intel_can_enable_sagv(struct drm_atomic_state *state);
> > -int intel_enable_sagv(struct drm_i915_private *dev_priv);
> > -int intel_disable_sagv(struct drm_i915_private *dev_priv);
> > +int intel_enable_sagv(struct drm_i915_private *i915);
> > +int intel_disable_sagv(struct drm_i915_private *i915);
> >   bool skl_wm_level_equals(const struct skl_wm_level *l1,
> >   			 const struct skl_wm_level *l2);
> >   bool skl_ddb_allocation_overlaps(const struct skl_ddb_entry *ddb,
> > @@ -67,24 +67,24 @@ void skl_write_cursor_wm(struct intel_plane *plane,
> >   bool ilk_disable_lp_wm(struct drm_device *dev);
> >   int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc,
> >   				  struct intel_crtc_state *cstate);
> > -void intel_init_ipc(struct drm_i915_private *dev_priv);
> > -void intel_enable_ipc(struct drm_i915_private *dev_priv);
> > +void intel_init_ipc(struct drm_i915_private *i915);
> > +void intel_enable_ipc(struct drm_i915_private *i915);
> > -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val);
> > -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val);
> > -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, i915_reg_t reg);
> > -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv, i915_reg_t reg);
> > +int intel_gpu_freq(struct drm_i915_private *i915, int val);
> > +int intel_freq_opcode(struct drm_i915_private *i915, int val);
> > +u64 intel_rc6_residency_ns(struct drm_i915_private *i915, i915_reg_t reg);
> > +u64 intel_rc6_residency_us(struct drm_i915_private *i915, i915_reg_t reg);
> > -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat1);
> > +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat1);
> > -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv);
> > -unsigned long i915_mch_val(struct drm_i915_private *dev_priv);
> > -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv);
> > -void i915_update_gfx_val(struct drm_i915_private *dev_priv);
> > +unsigned long i915_chipset_val(struct drm_i915_private *i915);
> > +unsigned long i915_mch_val(struct drm_i915_private *i915);
> > +unsigned long i915_gfx_val(struct drm_i915_private *i915);
> > +void i915_update_gfx_val(struct drm_i915_private *i915);
> > -bool ironlake_set_drps(struct drm_i915_private *dev_priv, u8 val);
> > -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val);
> > +bool ironlake_set_drps(struct drm_i915_private *i915, u8 val);
> > +int intel_set_rps(struct drm_i915_private *i915, u8 val);
> >   void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive);
> > -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable);
> > +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable);
> >   #endif /* __INTEL_PM_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_psr.c b/drivers/gpu/drm/i915/intel_psr.c
> > index 01ca502099df..072eea00b17a 100644
> > --- a/drivers/gpu/drm/i915/intel_psr.c
> > +++ b/drivers/gpu/drm/i915/intel_psr.c
> > @@ -71,14 +71,14 @@ static bool psr_global_enabled(u32 debug)
> >   	}
> >   }
> > -static bool intel_psr2_enabled(struct drm_i915_private *dev_priv,
> > +static bool intel_psr2_enabled(struct drm_i915_private *i915,
> >   			       const struct intel_crtc_state *crtc_state)
> >   {
> >   	/* Cannot enable DSC and PSR2 simultaneously */
> >   	WARN_ON(crtc_state->dsc_params.compression_enable &&
> >   		crtc_state->has_psr2);
> > -	switch (dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK) {
> > +	switch (i915->psr.debug & I915_PSR_DEBUG_MODE_MASK) {
> >   	case I915_PSR_DEBUG_DISABLE:
> >   	case I915_PSR_DEBUG_FORCE_PSR1:
> >   		return false;
> > @@ -104,20 +104,20 @@ static int edp_psr_shift(enum transcoder cpu_transcoder)
> >   	}
> >   }
> > -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug)
> > +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug)
> >   {
> >   	u32 debug_mask, mask;
> >   	enum transcoder cpu_transcoder;
> >   	u32 transcoders = BIT(TRANSCODER_EDP);
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		transcoders |= BIT(TRANSCODER_A) |
> >   			       BIT(TRANSCODER_B) |
> >   			       BIT(TRANSCODER_C);
> >   	debug_mask = 0;
> >   	mask = 0;
> > -	for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) {
> > +	for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) {
> >   		int shift = edp_psr_shift(cpu_transcoder);
> >   		mask |= EDP_PSR_ERROR(shift);
> > @@ -168,26 +168,26 @@ static void psr_event_print(u32 val, bool psr2_enabled)
> >   		DRM_DEBUG_KMS("\tPSR disabled\n");
> >   }
> > -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> > +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir)
> >   {
> >   	u32 transcoders = BIT(TRANSCODER_EDP);
> >   	enum transcoder cpu_transcoder;
> >   	ktime_t time_ns =  ktime_get();
> >   	u32 mask = 0;
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		transcoders |= BIT(TRANSCODER_A) |
> >   			       BIT(TRANSCODER_B) |
> >   			       BIT(TRANSCODER_C);
> > -	for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) {
> > +	for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) {
> >   		int shift = edp_psr_shift(cpu_transcoder);
> >   		if (psr_iir & EDP_PSR_ERROR(shift)) {
> >   			DRM_WARN("[transcoder %s] PSR aux error\n",
> >   				 transcoder_name(cpu_transcoder));
> > -			dev_priv->psr.irq_aux_error = true;
> > +			i915->psr.irq_aux_error = true;
> >   			/*
> >   			 * If this interruption is not masked it will keep
> > @@ -201,19 +201,19 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> >   		}
> >   		if (psr_iir & EDP_PSR_PRE_ENTRY(shift)) {
> > -			dev_priv->psr.last_entry_attempt = time_ns;
> > +			i915->psr.last_entry_attempt = time_ns;
> >   			DRM_DEBUG_KMS("[transcoder %s] PSR entry attempt in 2 vblanks\n",
> >   				      transcoder_name(cpu_transcoder));
> >   		}
> >   		if (psr_iir & EDP_PSR_POST_EXIT(shift)) {
> > -			dev_priv->psr.last_exit = time_ns;
> > +			i915->psr.last_exit = time_ns;
> >   			DRM_DEBUG_KMS("[transcoder %s] PSR exit completed\n",
> >   				      transcoder_name(cpu_transcoder));
> > -			if (INTEL_GEN(dev_priv) >= 9) {
> > +			if (INTEL_GEN(i915) >= 9) {
> >   				u32 val = I915_READ(PSR_EVENT(cpu_transcoder));
> > -				bool psr2_enabled = dev_priv->psr.psr2_enabled;
> > +				bool psr2_enabled = i915->psr.psr2_enabled;
> >   				I915_WRITE(PSR_EVENT(cpu_transcoder), val);
> >   				psr_event_print(val, psr2_enabled);
> > @@ -225,7 +225,7 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir)
> >   		mask |= I915_READ(EDP_PSR_IMR);
> >   		I915_WRITE(EDP_PSR_IMR, mask);
> > -		schedule_work(&dev_priv->psr.work);
> > +		schedule_work(&i915->psr.work);
> >   	}
> >   }
> > @@ -279,7 +279,7 @@ static u16 intel_dp_get_su_x_granulartiy(struct intel_dp *intel_dp)
> >   void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(dp_to_dig_port(intel_dp)->base.base.dev);
> >   	drm_dp_dpcd_read(&intel_dp->aux, DP_PSR_SUPPORT, intel_dp->psr_dpcd,
> > @@ -300,14 +300,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> >   		return;
> >   	}
> > -	dev_priv->psr.sink_support = true;
> > -	dev_priv->psr.sink_sync_latency =
> > +	i915->psr.sink_support = true;
> > +	i915->psr.sink_sync_latency =
> >   		intel_dp_get_sink_sync_latency(intel_dp);
> > -	WARN_ON(dev_priv->psr.dp);
> > -	dev_priv->psr.dp = intel_dp;
> > +	WARN_ON(i915->psr.dp);
> > +	i915->psr.dp = intel_dp;
> > -	if (INTEL_GEN(dev_priv) >= 9 &&
> > +	if (INTEL_GEN(i915) >= 9 &&
> >   	    (intel_dp->psr_dpcd[0] == DP_PSR2_WITH_Y_COORD_IS_SUPPORTED)) {
> >   		bool y_req = intel_dp->psr_dpcd[1] &
> >   			     DP_PSR2_SU_Y_COORDINATE_REQUIRED;
> > @@ -324,14 +324,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp)
> >   		 * Y-coordinate requirement panels we would need to enable
> >   		 * GTC first.
> >   		 */
> > -		dev_priv->psr.sink_psr2_support = y_req && alpm;
> > +		i915->psr.sink_psr2_support = y_req && alpm;
> >   		DRM_DEBUG_KMS("PSR2 %ssupported\n",
> > -			      dev_priv->psr.sink_psr2_support ? "" : "not ");
> > +			      i915->psr.sink_psr2_support ? "" : "not ");
> > -		if (dev_priv->psr.sink_psr2_support) {
> > -			dev_priv->psr.colorimetry_support =
> > +		if (i915->psr.sink_psr2_support) {
> > +			i915->psr.colorimetry_support =
> >   				intel_dp_get_colorimetry_status(intel_dp);
> > -			dev_priv->psr.su_x_granularity =
> > +			i915->psr.su_x_granularity =
> >   				intel_dp_get_su_x_granulartiy(intel_dp);
> >   		}
> >   	}
> > @@ -341,15 +341,15 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp,
> >   				const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	struct dp_sdp psr_vsc;
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		/* Prepare VSC Header for SU as per EDP 1.4 spec, Table 6.11 */
> >   		memset(&psr_vsc, 0, sizeof(psr_vsc));
> >   		psr_vsc.sdp_header.HB0 = 0;
> >   		psr_vsc.sdp_header.HB1 = 0x7;
> > -		if (dev_priv->psr.colorimetry_support) {
> > +		if (i915->psr.colorimetry_support) {
> >   			psr_vsc.sdp_header.HB2 = 0x5;
> >   			psr_vsc.sdp_header.HB3 = 0x13;
> >   		} else {
> > @@ -372,7 +372,7 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp,
> >   static void hsw_psr_setup_aux(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 aux_clock_divider, aux_ctl;
> >   	int i;
> >   	static const u8 aux_msg[] = {
> > @@ -405,19 +405,19 @@ static void hsw_psr_setup_aux(struct intel_dp *intel_dp)
> >   static void intel_psr_enable_sink(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u8 dpcd_val = DP_PSR_ENABLE;
> >   	/* Enable ALPM at sink for psr2 */
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		drm_dp_dpcd_writeb(&intel_dp->aux, DP_RECEIVER_ALPM_CONFIG,
> >   				   DP_ALPM_ENABLE);
> >   		dpcd_val |= DP_PSR_ENABLE_PSR2 | DP_PSR_IRQ_HPD_WITH_CRC_ERRORS;
> >   	} else {
> > -		if (dev_priv->psr.link_standby)
> > +		if (i915->psr.link_standby)
> >   			dpcd_val |= DP_PSR_MAIN_LINK_ACTIVE;
> > -		if (INTEL_GEN(dev_priv) >= 8)
> > +		if (INTEL_GEN(i915) >= 8)
> >   			dpcd_val |= DP_PSR_CRC_VERIFICATION;
> >   	}
> > @@ -428,26 +428,26 @@ static void intel_psr_enable_sink(struct intel_dp *intel_dp)
> >   static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 val = 0;
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		val |= EDP_PSR_TP4_TIME_0US;
> > -	if (dev_priv->vbt.psr.tp1_wakeup_time_us == 0)
> > +	if (i915->vbt.psr.tp1_wakeup_time_us == 0)
> >   		val |= EDP_PSR_TP1_TIME_0us;
> > -	else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 100)
> > +	else if (i915->vbt.psr.tp1_wakeup_time_us <= 100)
> >   		val |= EDP_PSR_TP1_TIME_100us;
> > -	else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 500)
> > +	else if (i915->vbt.psr.tp1_wakeup_time_us <= 500)
> >   		val |= EDP_PSR_TP1_TIME_500us;
> >   	else
> >   		val |= EDP_PSR_TP1_TIME_2500us;
> > -	if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us == 0)
> > +	if (i915->vbt.psr.tp2_tp3_wakeup_time_us == 0)
> >   		val |= EDP_PSR_TP2_TP3_TIME_0us;
> > -	else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 100)
> > +	else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 100)
> >   		val |= EDP_PSR_TP2_TP3_TIME_100us;
> > -	else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 500)
> > +	else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 500)
> >   		val |= EDP_PSR_TP2_TP3_TIME_500us;
> >   	else
> >   		val |= EDP_PSR_TP2_TP3_TIME_2500us;
> > @@ -463,31 +463,31 @@ static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp)
> >   static void hsw_activate_psr1(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 max_sleep_time = 0x1f;
> >   	u32 val = EDP_PSR_ENABLE;
> >   	/* Let's use 6 as the minimum to cover all known cases including the
> >   	 * off-by-one issue that HW has in some cases.
> >   	 */
> > -	int idle_frames = max(6, dev_priv->vbt.psr.idle_frames);
> > +	int idle_frames = max(6, i915->vbt.psr.idle_frames);
> >   	/* sink_sync_latency of 8 means source has to wait for more than 8
> >   	 * frames, we'll go with 9 frames for now
> >   	 */
> > -	idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1);
> > +	idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1);
> >   	val |= idle_frames << EDP_PSR_IDLE_FRAME_SHIFT;
> >   	val |= max_sleep_time << EDP_PSR_MAX_SLEEP_TIME_SHIFT;
> > -	if (IS_HASWELL(dev_priv))
> > +	if (IS_HASWELL(i915))
> >   		val |= EDP_PSR_MIN_LINK_ENTRY_TIME_8_LINES;
> > -	if (dev_priv->psr.link_standby)
> > +	if (i915->psr.link_standby)
> >   		val |= EDP_PSR_LINK_STANDBY;
> >   	val |= intel_psr1_get_tp_time(intel_dp);
> > -	if (INTEL_GEN(dev_priv) >= 8)
> > +	if (INTEL_GEN(i915) >= 8)
> >   		val |= EDP_PSR_CRC_ENABLE;
> >   	val |= I915_READ(EDP_PSR_CTL) & EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK;
> > @@ -496,29 +496,29 @@ static void hsw_activate_psr1(struct intel_dp *intel_dp)
> >   static void hsw_activate_psr2(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	u32 val;
> >   	/* Let's use 6 as the minimum to cover all known cases including the
> >   	 * off-by-one issue that HW has in some cases.
> >   	 */
> > -	int idle_frames = max(6, dev_priv->vbt.psr.idle_frames);
> > +	int idle_frames = max(6, i915->vbt.psr.idle_frames);
> > -	idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1);
> > +	idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1);
> >   	val = idle_frames << EDP_PSR2_IDLE_FRAME_SHIFT;
> >   	val |= EDP_PSR2_ENABLE | EDP_SU_TRACK_ENABLE;
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		val |= EDP_Y_COORDINATE_ENABLE;
> > -	val |= EDP_PSR2_FRAME_BEFORE_SU(dev_priv->psr.sink_sync_latency + 1);
> > +	val |= EDP_PSR2_FRAME_BEFORE_SU(i915->psr.sink_sync_latency + 1);
> > -	if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 &&
> > -	    dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50)
> > +	if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 &&
> > +	    i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50)
> >   		val |= EDP_PSR2_TP2_TIME_50us;
> > -	else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100)
> > +	else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100)
> >   		val |= EDP_PSR2_TP2_TIME_100us;
> > -	else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500)
> > +	else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500)
> >   		val |= EDP_PSR2_TP2_TIME_500us;
> >   	else
> >   		val |= EDP_PSR2_TP2_TIME_2500us;
> > @@ -535,12 +535,12 @@ static void hsw_activate_psr2(struct intel_dp *intel_dp)
> >   static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> >   				    struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	int crtc_hdisplay = crtc_state->base.adjusted_mode.crtc_hdisplay;
> >   	int crtc_vdisplay = crtc_state->base.adjusted_mode.crtc_vdisplay;
> >   	int psr_max_h = 0, psr_max_v = 0;
> > -	if (!dev_priv->psr.sink_psr2_support)
> > +	if (!i915->psr.sink_psr2_support)
> >   		return false;
> >   	/*
> > @@ -553,10 +553,10 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> >   		return false;
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) {
> >   		psr_max_h = 4096;
> >   		psr_max_v = 2304;
> > -	} else if (IS_GEN(dev_priv, 9)) {
> > +	} else if (IS_GEN(i915, 9)) {
> >   		psr_max_h = 3640;
> >   		psr_max_v = 2304;
> >   	}
> > @@ -574,9 +574,9 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp,
> >   	 * only need to validate the SU block width is a multiple of
> >   	 * x granularity.
> >   	 */
> > -	if (crtc_hdisplay % dev_priv->psr.su_x_granularity) {
> > +	if (crtc_hdisplay % i915->psr.su_x_granularity) {
> >   		DRM_DEBUG_KMS("PSR2 not enabled, hdisplay(%d) not multiple of %d\n",
> > -			      crtc_hdisplay, dev_priv->psr.su_x_granularity);
> > +			      crtc_hdisplay, i915->psr.su_x_granularity);
> >   		return false;
> >   	}
> > @@ -592,15 +592,15 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
> >   			      struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp);
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	const struct drm_display_mode *adjusted_mode =
> >   		&crtc_state->base.adjusted_mode;
> >   	int psr_setup_time;
> > -	if (!CAN_PSR(dev_priv))
> > +	if (!CAN_PSR(i915))
> >   		return;
> > -	if (intel_dp != dev_priv->psr.dp)
> > +	if (intel_dp != i915->psr.dp)
> >   		return;
> >   	/*
> > @@ -615,7 +615,7 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
> >   		return;
> >   	}
> > -	if (dev_priv->psr.sink_not_reliable) {
> > +	if (i915->psr.sink_not_reliable) {
> >   		DRM_DEBUG_KMS("PSR sink implementation is not reliable\n");
> >   		return;
> >   	}
> > @@ -645,24 +645,24 @@ void intel_psr_compute_config(struct intel_dp *intel_dp,
> >   static void intel_psr_activate(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > +	if (INTEL_GEN(i915) >= 9)
> >   		WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE);
> >   	WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE);
> > -	WARN_ON(dev_priv->psr.active);
> > -	lockdep_assert_held(&dev_priv->psr.lock);
> > +	WARN_ON(i915->psr.active);
> > +	lockdep_assert_held(&i915->psr.lock);
> >   	/* psr1 and psr2 are mutually exclusive.*/
> > -	if (dev_priv->psr.psr2_enabled)
> > +	if (i915->psr.psr2_enabled)
> >   		hsw_activate_psr2(intel_dp);
> >   	else
> >   		hsw_activate_psr1(intel_dp);
> > -	dev_priv->psr.active = true;
> > +	i915->psr.active = true;
> >   }
> > -static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> > +static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *i915,
> >   					 enum transcoder cpu_transcoder)
> >   {
> >   	static const i915_reg_t regs[] = {
> > @@ -672,7 +672,7 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> >   		[TRANSCODER_EDP] = CHICKEN_TRANS_EDP,
> >   	};
> > -	WARN_ON(INTEL_GEN(dev_priv) < 9);
> > +	WARN_ON(INTEL_GEN(i915) < 9);
> >   	if (WARN_ON(cpu_transcoder >= ARRAY_SIZE(regs) ||
> >   		    !regs[cpu_transcoder].reg))
> > @@ -684,19 +684,19 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv,
> >   static void intel_psr_enable_source(struct intel_dp *intel_dp,
> >   				    const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> >   	u32 mask;
> >   	/* Only HSW and BDW have PSR AUX registers that need to be setup. SKL+
> >   	 * use hardcoded values PSR AUX transactions
> >   	 */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		hsw_psr_setup_aux(intel_dp);
> > -	if (dev_priv->psr.psr2_enabled && (IS_GEN(dev_priv, 9) &&
> > -					   !IS_GEMINILAKE(dev_priv))) {
> > -		i915_reg_t reg = gen9_chicken_trans_reg(dev_priv,
> > +	if (i915->psr.psr2_enabled && (IS_GEN(i915, 9) &&
> > +					   !IS_GEMINILAKE(i915))) {
> > +		i915_reg_t reg = gen9_chicken_trans_reg(i915,
> >   							cpu_transcoder);
> >   		u32 chicken = I915_READ(reg);
> > @@ -716,29 +716,29 @@ static void intel_psr_enable_source(struct intel_dp *intel_dp,
> >   	       EDP_PSR_DEBUG_MASK_LPSP |
> >   	       EDP_PSR_DEBUG_MASK_MAX_SLEEP;
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		mask |= EDP_PSR_DEBUG_MASK_DISP_REG_WRITE;
> >   	I915_WRITE(EDP_PSR_DEBUG, mask);
> >   }
> > -static void intel_psr_enable_locked(struct drm_i915_private *dev_priv,
> > +static void intel_psr_enable_locked(struct drm_i915_private *i915,
> >   				    const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct intel_dp *intel_dp = dev_priv->psr.dp;
> > +	struct intel_dp *intel_dp = i915->psr.dp;
> > -	WARN_ON(dev_priv->psr.enabled);
> > +	WARN_ON(i915->psr.enabled);
> > -	dev_priv->psr.psr2_enabled = intel_psr2_enabled(dev_priv, crtc_state);
> > -	dev_priv->psr.busy_frontbuffer_bits = 0;
> > -	dev_priv->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> > +	i915->psr.psr2_enabled = intel_psr2_enabled(i915, crtc_state);
> > +	i915->psr.busy_frontbuffer_bits = 0;
> > +	i915->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe;
> >   	DRM_DEBUG_KMS("Enabling PSR%s\n",
> > -		      dev_priv->psr.psr2_enabled ? "2" : "1");
> > +		      i915->psr.psr2_enabled ? "2" : "1");
> >   	intel_psr_setup_vsc(intel_dp, crtc_state);
> >   	intel_psr_enable_sink(intel_dp);
> >   	intel_psr_enable_source(intel_dp, crtc_state);
> > -	dev_priv->psr.enabled = true;
> > +	i915->psr.enabled = true;
> >   	intel_psr_activate(intel_dp);
> >   }
> > @@ -753,41 +753,41 @@ static void intel_psr_enable_locked(struct drm_i915_private *dev_priv,
> >   void intel_psr_enable(struct intel_dp *intel_dp,
> >   		      const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (!crtc_state->has_psr)
> >   		return;
> > -	if (WARN_ON(!CAN_PSR(dev_priv)))
> > +	if (WARN_ON(!CAN_PSR(i915)))
> >   		return;
> > -	WARN_ON(dev_priv->drrs.dp);
> > +	WARN_ON(i915->drrs.dp);
> > -	mutex_lock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> > -	if (!psr_global_enabled(dev_priv->psr.debug)) {
> > +	if (!psr_global_enabled(i915->psr.debug)) {
> >   		DRM_DEBUG_KMS("PSR disabled by flag\n");
> >   		goto unlock;
> >   	}
> > -	intel_psr_enable_locked(dev_priv, crtc_state);
> > +	intel_psr_enable_locked(i915, crtc_state);
> >   unlock:
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> >   }
> > -static void intel_psr_exit(struct drm_i915_private *dev_priv)
> > +static void intel_psr_exit(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	if (!dev_priv->psr.active) {
> > -		if (INTEL_GEN(dev_priv) >= 9)
> > +	if (!i915->psr.active) {
> > +		if (INTEL_GEN(i915) >= 9)
> >   			WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE);
> >   		WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE);
> >   		return;
> >   	}
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		val = I915_READ(EDP_PSR2_CTL);
> >   		WARN_ON(!(val & EDP_PSR2_ENABLE));
> >   		I915_WRITE(EDP_PSR2_CTL, val & ~EDP_PSR2_ENABLE);
> > @@ -796,26 +796,26 @@ static void intel_psr_exit(struct drm_i915_private *dev_priv)
> >   		WARN_ON(!(val & EDP_PSR_ENABLE));
> >   		I915_WRITE(EDP_PSR_CTL, val & ~EDP_PSR_ENABLE);
> >   	}
> > -	dev_priv->psr.active = false;
> > +	i915->psr.active = false;
> >   }
> >   static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	i915_reg_t psr_status;
> >   	u32 psr_status_mask;
> > -	lockdep_assert_held(&dev_priv->psr.lock);
> > +	lockdep_assert_held(&i915->psr.lock);
> > -	if (!dev_priv->psr.enabled)
> > +	if (!i915->psr.enabled)
> >   		return;
> >   	DRM_DEBUG_KMS("Disabling PSR%s\n",
> > -		      dev_priv->psr.psr2_enabled ? "2" : "1");
> > +		      i915->psr.psr2_enabled ? "2" : "1");
> > -	intel_psr_exit(dev_priv);
> > +	intel_psr_exit(i915);
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		psr_status = EDP_PSR2_STATUS;
> >   		psr_status_mask = EDP_PSR2_STATUS_STATE_MASK;
> >   	} else {
> > @@ -824,14 +824,14 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> >   	}
> >   	/* Wait till PSR is idle */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    psr_status, psr_status_mask, 0, 2000))
> >   		DRM_ERROR("Timed out waiting PSR idle state\n");
> >   	/* Disable PSR on Sink */
> >   	drm_dp_dpcd_writeb(&intel_dp->aux, DP_PSR_EN_CFG, 0);
> > -	dev_priv->psr.enabled = false;
> > +	i915->psr.enabled = false;
> >   }
> >   /**
> > @@ -844,23 +844,23 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp)
> >   void intel_psr_disable(struct intel_dp *intel_dp,
> >   		       const struct intel_crtc_state *old_crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	if (!old_crtc_state->has_psr)
> >   		return;
> > -	if (WARN_ON(!CAN_PSR(dev_priv)))
> > +	if (WARN_ON(!CAN_PSR(i915)))
> >   		return;
> > -	mutex_lock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> >   	intel_psr_disable_locked(intel_dp);
> > -	mutex_unlock(&dev_priv->psr.lock);
> > -	cancel_work_sync(&dev_priv->psr.work);
> > +	mutex_unlock(&i915->psr.lock);
> > +	cancel_work_sync(&i915->psr.work);
> >   }
> > -static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> > +static void psr_force_hw_tracking_exit(struct drm_i915_private *i915)
> >   {
> >   	/*
> >   	 * Display WA #0884: all
> > @@ -871,7 +871,7 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> >   	 * but it makes more sense write to the current active
> >   	 * pipe.
> >   	 */
> > -	I915_WRITE(CURSURFLIVE(dev_priv->psr.pipe), 0);
> > +	I915_WRITE(CURSURFLIVE(i915->psr.pipe), 0);
> >   }
> >   /**
> > @@ -886,22 +886,22 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv)
> >   void intel_psr_update(struct intel_dp *intel_dp,
> >   		      const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > -	struct i915_psr *psr = &dev_priv->psr;
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > +	struct i915_psr *psr = &i915->psr;
> >   	bool enable, psr2_enable;
> > -	if (!CAN_PSR(dev_priv) || READ_ONCE(psr->dp) != intel_dp)
> > +	if (!CAN_PSR(i915) || READ_ONCE(psr->dp) != intel_dp)
> >   		return;
> > -	mutex_lock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> >   	enable = crtc_state->has_psr && psr_global_enabled(psr->debug);
> > -	psr2_enable = intel_psr2_enabled(dev_priv, crtc_state);
> > +	psr2_enable = intel_psr2_enabled(i915, crtc_state);
> >   	if (enable == psr->enabled && psr2_enable == psr->psr2_enabled) {
> >   		/* Force a PSR exit when enabling CRC to avoid CRC timeouts */
> >   		if (crtc_state->crc_enabled && psr->enabled)
> > -			psr_force_hw_tracking_exit(dev_priv);
> > +			psr_force_hw_tracking_exit(i915);
> >   		goto unlock;
> >   	}
> > @@ -910,10 +910,10 @@ void intel_psr_update(struct intel_dp *intel_dp,
> >   		intel_psr_disable_locked(intel_dp);
> >   	if (enable)
> > -		intel_psr_enable_locked(dev_priv, crtc_state);
> > +		intel_psr_enable_locked(i915, crtc_state);
> >   unlock:
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> >   }
> >   /**
> > @@ -930,13 +930,13 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> >   			    u32 *out_value)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> > -	if (!dev_priv->psr.enabled || !new_crtc_state->has_psr)
> > +	if (!i915->psr.enabled || !new_crtc_state->has_psr)
> >   		return 0;
> >   	/* FIXME: Update this for PSR2 if we need to wait for idle */
> > -	if (READ_ONCE(dev_priv->psr.psr2_enabled))
> > +	if (READ_ONCE(i915->psr.psr2_enabled))
> >   		return 0;
> >   	/*
> > @@ -946,22 +946,22 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> >   	 * defensive enough to cover everything.
> >   	 */
> > -	return __intel_wait_for_register(&dev_priv->uncore, EDP_PSR_STATUS,
> > +	return __intel_wait_for_register(&i915->uncore, EDP_PSR_STATUS,
> >   					 EDP_PSR_STATUS_STATE_MASK,
> >   					 EDP_PSR_STATUS_STATE_IDLE, 2, 50,
> >   					 out_value);
> >   }
> > -static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv)
> > +static bool __psr_wait_for_idle_locked(struct drm_i915_private *i915)
> >   {
> >   	i915_reg_t reg;
> >   	u32 mask;
> >   	int err;
> > -	if (!dev_priv->psr.enabled)
> > +	if (!i915->psr.enabled)
> >   		return false;
> > -	if (dev_priv->psr.psr2_enabled) {
> > +	if (i915->psr.psr2_enabled) {
> >   		reg = EDP_PSR2_STATUS;
> >   		mask = EDP_PSR2_STATUS_STATE_MASK;
> >   	} else {
> > @@ -969,20 +969,20 @@ static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv)
> >   		mask = EDP_PSR_STATUS_STATE_MASK;
> >   	}
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> > -	err = intel_wait_for_register(&dev_priv->uncore, reg, mask, 0, 50);
> > +	err = intel_wait_for_register(&i915->uncore, reg, mask, 0, 50);
> >   	if (err)
> >   		DRM_ERROR("Timed out waiting for PSR Idle for re-enable\n");
> >   	/* After the unlocked wait, verify that PSR is still wanted! */
> > -	mutex_lock(&dev_priv->psr.lock);
> > -	return err == 0 && dev_priv->psr.enabled;
> > +	mutex_lock(&i915->psr.lock);
> > +	return err == 0 && i915->psr.enabled;
> >   }
> > -static int intel_psr_fastset_force(struct drm_i915_private *dev_priv)
> > +static int intel_psr_fastset_force(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_modeset_acquire_ctx ctx;
> >   	struct drm_atomic_state *state;
> >   	struct drm_crtc *crtc;
> > @@ -1032,7 +1032,7 @@ static int intel_psr_fastset_force(struct drm_i915_private *dev_priv)
> >   	return err;
> >   }
> > -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val)
> > +int intel_psr_debug_set(struct drm_i915_private *i915, u64 val)
> >   {
> >   	const u32 mode = val & I915_PSR_DEBUG_MODE_MASK;
> >   	u32 old_mode;
> > @@ -1044,25 +1044,25 @@ int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val)
> >   		return -EINVAL;
> >   	}
> > -	ret = mutex_lock_interruptible(&dev_priv->psr.lock);
> > +	ret = mutex_lock_interruptible(&i915->psr.lock);
> >   	if (ret)
> >   		return ret;
> > -	old_mode = dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK;
> > -	dev_priv->psr.debug = val;
> > -	intel_psr_irq_control(dev_priv, dev_priv->psr.debug);
> > +	old_mode = i915->psr.debug & I915_PSR_DEBUG_MODE_MASK;
> > +	i915->psr.debug = val;
> > +	intel_psr_irq_control(i915, i915->psr.debug);
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> >   	if (old_mode != mode)
> > -		ret = intel_psr_fastset_force(dev_priv);
> > +		ret = intel_psr_fastset_force(i915);
> >   	return ret;
> >   }
> > -static void intel_psr_handle_irq(struct drm_i915_private *dev_priv)
> > +static void intel_psr_handle_irq(struct drm_i915_private *i915)
> >   {
> > -	struct i915_psr *psr = &dev_priv->psr;
> > +	struct i915_psr *psr = &i915->psr;
> >   	intel_psr_disable_locked(psr->dp);
> >   	psr->sink_not_reliable = true;
> > @@ -1072,16 +1072,16 @@ static void intel_psr_handle_irq(struct drm_i915_private *dev_priv)
> >   static void intel_psr_work(struct work_struct *work)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > -		container_of(work, typeof(*dev_priv), psr.work);
> > +	struct drm_i915_private *i915 =
> > +		container_of(work, typeof(*i915), psr.work);
> > -	mutex_lock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> > -	if (!dev_priv->psr.enabled)
> > +	if (!i915->psr.enabled)
> >   		goto unlock;
> > -	if (READ_ONCE(dev_priv->psr.irq_aux_error))
> > -		intel_psr_handle_irq(dev_priv);
> > +	if (READ_ONCE(i915->psr.irq_aux_error))
> > +		intel_psr_handle_irq(i915);
> >   	/*
> >   	 * We have to make sure PSR is ready for re-enable
> > @@ -1089,7 +1089,7 @@ static void intel_psr_work(struct work_struct *work)
> >   	 * PSR might take some time to get fully disabled
> >   	 * and be ready for re-enable.
> >   	 */
> > -	if (!__psr_wait_for_idle_locked(dev_priv))
> > +	if (!__psr_wait_for_idle_locked(i915))
> >   		goto unlock;
> >   	/*
> > @@ -1097,17 +1097,17 @@ static void intel_psr_work(struct work_struct *work)
> >   	 * recheck. Since psr_flush first clears this and then reschedules we
> >   	 * won't ever miss a flush when bailing out here.
> >   	 */
> > -	if (dev_priv->psr.busy_frontbuffer_bits || dev_priv->psr.active)
> > +	if (i915->psr.busy_frontbuffer_bits || i915->psr.active)
> >   		goto unlock;
> > -	intel_psr_activate(dev_priv->psr.dp);
> > +	intel_psr_activate(i915->psr.dp);
> >   unlock:
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> >   }
> >   /**
> >    * intel_psr_invalidate - Invalidade PSR
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    * @origin: which operation caused the invalidate
> >    *
> > @@ -1118,33 +1118,33 @@ static void intel_psr_work(struct work_struct *work)
> >    *
> >    * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits."
> >    */
> > -void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> > +void intel_psr_invalidate(struct drm_i915_private *i915,
> >   			  unsigned frontbuffer_bits, enum fb_op_origin origin)
> >   {
> > -	if (!CAN_PSR(dev_priv))
> > +	if (!CAN_PSR(i915))
> >   		return;
> >   	if (origin == ORIGIN_FLIP)
> >   		return;
> > -	mutex_lock(&dev_priv->psr.lock);
> > -	if (!dev_priv->psr.enabled) {
> > -		mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> > +	if (!i915->psr.enabled) {
> > +		mutex_unlock(&i915->psr.lock);
> >   		return;
> >   	}
> > -	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe);
> > -	dev_priv->psr.busy_frontbuffer_bits |= frontbuffer_bits;
> > +	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe);
> > +	i915->psr.busy_frontbuffer_bits |= frontbuffer_bits;
> >   	if (frontbuffer_bits)
> > -		intel_psr_exit(dev_priv);
> > +		intel_psr_exit(i915);
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_unlock(&i915->psr.lock);
> >   }
> >   /**
> >    * intel_psr_flush - Flush PSR
> > - * @dev_priv: i915 device
> > + * @i915: i915 device
> >    * @frontbuffer_bits: frontbuffer plane tracking bits
> >    * @origin: which operation caused the flush
> >    *
> > @@ -1155,55 +1155,55 @@ void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> >    *
> >    * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits.
> >    */
> > -void intel_psr_flush(struct drm_i915_private *dev_priv,
> > +void intel_psr_flush(struct drm_i915_private *i915,
> >   		     unsigned frontbuffer_bits, enum fb_op_origin origin)
> >   {
> > -	if (!CAN_PSR(dev_priv))
> > +	if (!CAN_PSR(i915))
> >   		return;
> >   	if (origin == ORIGIN_FLIP)
> >   		return;
> > -	mutex_lock(&dev_priv->psr.lock);
> > -	if (!dev_priv->psr.enabled) {
> > -		mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> > +	if (!i915->psr.enabled) {
> > +		mutex_unlock(&i915->psr.lock);
> >   		return;
> >   	}
> > -	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe);
> > -	dev_priv->psr.busy_frontbuffer_bits &= ~frontbuffer_bits;
> > +	frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe);
> > +	i915->psr.busy_frontbuffer_bits &= ~frontbuffer_bits;
> >   	/* By definition flush = invalidate + flush */
> >   	if (frontbuffer_bits)
> > -		psr_force_hw_tracking_exit(dev_priv);
> > +		psr_force_hw_tracking_exit(i915);
> > -	if (!dev_priv->psr.active && !dev_priv->psr.busy_frontbuffer_bits)
> > -		schedule_work(&dev_priv->psr.work);
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	if (!i915->psr.active && !i915->psr.busy_frontbuffer_bits)
> > +		schedule_work(&i915->psr.work);
> > +	mutex_unlock(&i915->psr.lock);
> >   }
> >   /**
> >    * intel_psr_init - Init basic PSR work and mutex.
> > - * @dev_priv: i915 device private
> > + * @i915: i915 device private
> >    *
> >    * This function is  called only once at driver load to initialize basic
> >    * PSR stuff.
> >    */
> > -void intel_psr_init(struct drm_i915_private *dev_priv)
> > +void intel_psr_init(struct drm_i915_private *i915)
> >   {
> >   	u32 val;
> > -	if (!HAS_PSR(dev_priv))
> > +	if (!HAS_PSR(i915))
> >   		return;
> > -	dev_priv->psr_mmio_base = IS_HASWELL(dev_priv) ?
> > +	i915->psr_mmio_base = IS_HASWELL(i915) ?
> >   		HSW_EDP_PSR_BASE : BDW_EDP_PSR_BASE;
> > -	if (!dev_priv->psr.sink_support)
> > +	if (!i915->psr.sink_support)
> >   		return;
> >   	if (i915_modparams.enable_psr == -1)
> > -		if (INTEL_GEN(dev_priv) < 9 || !dev_priv->vbt.psr.enable)
> > +		if (INTEL_GEN(i915) < 9 || !i915->vbt.psr.enable)
> >   			i915_modparams.enable_psr = 0;
> >   	/*
> > @@ -1218,31 +1218,31 @@ void intel_psr_init(struct drm_i915_private *dev_priv)
> >   	val &= EDP_PSR_ERROR(edp_psr_shift(TRANSCODER_EDP));
> >   	if (val) {
> >   		DRM_DEBUG_KMS("PSR interruption error set\n");
> > -		dev_priv->psr.sink_not_reliable = true;
> > +		i915->psr.sink_not_reliable = true;
> >   	}
> >   	/* Set link_standby x link_off defaults */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915))
> >   		/* HSW and BDW require workarounds that we don't implement. */
> > -		dev_priv->psr.link_standby = false;
> > +		i915->psr.link_standby = false;
> >   	else
> >   		/* For new platforms let's respect VBT back again */
> > -		dev_priv->psr.link_standby = dev_priv->vbt.psr.full_link;
> > +		i915->psr.link_standby = i915->vbt.psr.full_link;
> > -	INIT_WORK(&dev_priv->psr.work, intel_psr_work);
> > -	mutex_init(&dev_priv->psr.lock);
> > +	INIT_WORK(&i915->psr.work, intel_psr_work);
> > +	mutex_init(&i915->psr.lock);
> >   }
> >   void intel_psr_short_pulse(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > -	struct i915_psr *psr = &dev_priv->psr;
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> > +	struct i915_psr *psr = &i915->psr;
> >   	u8 val;
> >   	const u8 errors = DP_PSR_RFB_STORAGE_ERROR |
> >   			  DP_PSR_VSC_SDP_UNCORRECTABLE_ERROR |
> >   			  DP_PSR_LINK_CRC_ERROR;
> > -	if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp))
> > +	if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp))
> >   		return;
> >   	mutex_lock(&psr->lock);
> > @@ -1288,15 +1288,15 @@ void intel_psr_short_pulse(struct intel_dp *intel_dp)
> >   bool intel_psr_enabled(struct intel_dp *intel_dp)
> >   {
> > -	struct drm_i915_private *dev_priv = dp_to_i915(intel_dp);
> > +	struct drm_i915_private *i915 = dp_to_i915(intel_dp);
> >   	bool ret;
> > -	if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp))
> > +	if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp))
> >   		return false;
> > -	mutex_lock(&dev_priv->psr.lock);
> > -	ret = (dev_priv->psr.dp == intel_dp && dev_priv->psr.enabled);
> > -	mutex_unlock(&dev_priv->psr.lock);
> > +	mutex_lock(&i915->psr.lock);
> > +	ret = (i915->psr.dp == intel_dp && i915->psr.enabled);
> > +	mutex_unlock(&i915->psr.lock);
> >   	return ret;
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_psr.h b/drivers/gpu/drm/i915/intel_psr.h
> > index dc818826f36d..8a3ddc4ae44d 100644
> > --- a/drivers/gpu/drm/i915/intel_psr.h
> > +++ b/drivers/gpu/drm/i915/intel_psr.h
> > @@ -12,7 +12,7 @@ struct drm_i915_private;
> >   struct intel_crtc_state;
> >   struct intel_dp;
> > -#define CAN_PSR(dev_priv) (HAS_PSR(dev_priv) && dev_priv->psr.sink_support)
> > +#define CAN_PSR(i915) (HAS_PSR(i915) && i915->psr.sink_support)
> >   void intel_psr_init_dpcd(struct intel_dp *intel_dp);
> >   void intel_psr_enable(struct intel_dp *intel_dp,
> >   		      const struct intel_crtc_state *crtc_state);
> > @@ -20,18 +20,18 @@ void intel_psr_disable(struct intel_dp *intel_dp,
> >   		       const struct intel_crtc_state *old_crtc_state);
> >   void intel_psr_update(struct intel_dp *intel_dp,
> >   		      const struct intel_crtc_state *crtc_state);
> > -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 value);
> > -void intel_psr_invalidate(struct drm_i915_private *dev_priv,
> > +int intel_psr_debug_set(struct drm_i915_private *i915, u64 value);
> > +void intel_psr_invalidate(struct drm_i915_private *i915,
> >   			  unsigned frontbuffer_bits,
> >   			  enum fb_op_origin origin);
> > -void intel_psr_flush(struct drm_i915_private *dev_priv,
> > +void intel_psr_flush(struct drm_i915_private *i915,
> >   		     unsigned frontbuffer_bits,
> >   		     enum fb_op_origin origin);
> > -void intel_psr_init(struct drm_i915_private *dev_priv);
> > +void intel_psr_init(struct drm_i915_private *i915);
> >   void intel_psr_compute_config(struct intel_dp *intel_dp,
> >   			      struct intel_crtc_state *crtc_state);
> > -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug);
> > -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir);
> > +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug);
> > +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir);
> >   void intel_psr_short_pulse(struct intel_dp *intel_dp);
> >   int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state,
> >   			    u32 *out_value);
> > diff --git a/drivers/gpu/drm/i915/intel_quirks.h b/drivers/gpu/drm/i915/intel_quirks.h
> > index b0fcff142a56..b617ca047695 100644
> > --- a/drivers/gpu/drm/i915/intel_quirks.h
> > +++ b/drivers/gpu/drm/i915/intel_quirks.h
> > @@ -8,6 +8,6 @@
> >   struct drm_i915_private;
> > -void intel_init_quirks(struct drm_i915_private *dev_priv);
> > +void intel_init_quirks(struct drm_i915_private *i915);
> >   #endif /* __INTEL_QUIRKS_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_runtime_pm.h b/drivers/gpu/drm/i915/intel_runtime_pm.h
> > index a7acceb13473..1042f1c0b125 100644
> > --- a/drivers/gpu/drm/i915/intel_runtime_pm.h
> > +++ b/drivers/gpu/drm/i915/intel_runtime_pm.h
> > @@ -20,10 +20,10 @@ enum i915_drm_suspend_mode {
> >   	I915_DRM_SUSPEND_HIBERNATE,
> >   };
> > -void intel_runtime_pm_init_early(struct drm_i915_private *dev_priv);
> > -void intel_runtime_pm_enable(struct drm_i915_private *dev_priv);
> > -void intel_runtime_pm_disable(struct drm_i915_private *dev_priv);
> > -void intel_runtime_pm_cleanup(struct drm_i915_private *dev_priv);
> > +void intel_runtime_pm_init_early(struct drm_i915_private *i915);
> > +void intel_runtime_pm_enable(struct drm_i915_private *i915);
> > +void intel_runtime_pm_disable(struct drm_i915_private *i915);
> > +void intel_runtime_pm_cleanup(struct drm_i915_private *i915);
> >   intel_wakeref_t intel_runtime_pm_get(struct drm_i915_private *i915);
> >   intel_wakeref_t intel_runtime_pm_get_if_in_use(struct drm_i915_private *i915);
> > diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c
> > index 0860ae36bb87..b916c79d6b31 100644
> > --- a/drivers/gpu/drm/i915/intel_sdvo.c
> > +++ b/drivers/gpu/drm/i915/intel_sdvo.c
> > @@ -212,18 +212,18 @@ intel_sdvo_create_enhance_property(struct intel_sdvo *intel_sdvo,
> >   static void intel_sdvo_write_sdvox(struct intel_sdvo *intel_sdvo, u32 val)
> >   {
> >   	struct drm_device *dev = intel_sdvo->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 bval = val, cval = val;
> >   	int i;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		I915_WRITE(intel_sdvo->sdvo_reg, val);
> >   		POSTING_READ(intel_sdvo->sdvo_reg);
> >   		/*
> >   		 * HW workaround, need to write this twice for issue
> >   		 * that may result in first write getting masked.
> >   		 */
> > -		if (HAS_PCH_IBX(dev_priv)) {
> > +		if (HAS_PCH_IBX(i915)) {
> >   			I915_WRITE(intel_sdvo->sdvo_reg, val);
> >   			POSTING_READ(intel_sdvo->sdvo_reg);
> >   		}
> > @@ -1400,7 +1400,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> >   				  const struct intel_crtc_state *crtc_state,
> >   				  const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode;
> >   	const struct intel_sdvo_connector_state *sdvo_state =
> > @@ -1481,13 +1481,13 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> >   		return;
> >   	/* Set the SDVO control regs. */
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		/* The real mode polarity is set by the SDVO commands, using
> >   		 * struct intel_sdvo_dtd. */
> >   		sdvox = SDVO_VSYNC_ACTIVE_HIGH | SDVO_HSYNC_ACTIVE_HIGH;
> > -		if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range)
> > +		if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range)
> >   			sdvox |= HDMI_COLOR_RANGE_16_235;
> > -		if (INTEL_GEN(dev_priv) < 5)
> > +		if (INTEL_GEN(i915) < 5)
> >   			sdvox |= SDVO_BORDER_ENABLE;
> >   	} else {
> >   		sdvox = I915_READ(intel_sdvo->sdvo_reg);
> > @@ -1498,15 +1498,15 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> >   		sdvox |= (9 << 19) | SDVO_BORDER_ENABLE;
> >   	}
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		sdvox |= SDVO_PIPE_SEL_CPT(crtc->pipe);
> >   	else
> >   		sdvox |= SDVO_PIPE_SEL(crtc->pipe);
> > -	if (INTEL_GEN(dev_priv) >= 4) {
> > +	if (INTEL_GEN(i915) >= 4) {
> >   		/* done in crtc_mode_set as the dpll_md reg must be written early */
> > -	} else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) ||
> > -		   IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) {
> > +	} else if (IS_I945G(i915) || IS_I945GM(i915) ||
> > +		   IS_G33(i915) || IS_PINEVIEW(i915)) {
> >   		/* done in crtc_mode_set as it lives inside the dpll register */
> >   	} else {
> >   		sdvox |= (crtc_state->pixel_multiplier - 1)
> > @@ -1514,7 +1514,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder,
> >   	}
> >   	if (input_dtd.part2.sdvo_flags & SDVO_NEED_TO_STALL &&
> > -	    INTEL_GEN(dev_priv) < 5)
> > +	    INTEL_GEN(i915) < 5)
> >   		sdvox |= SDVO_STALL_SELECT;
> >   	intel_sdvo_write_sdvox(intel_sdvo, sdvox);
> >   }
> > @@ -1531,7 +1531,7 @@ static bool intel_sdvo_connector_get_hw_state(struct intel_connector *connector)
> >   	return active_outputs & intel_sdvo_connector->output_flag;
> >   }
> > -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_sdvo_port_enabled(struct drm_i915_private *i915,
> >   			     i915_reg_t sdvo_reg, enum pipe *pipe)
> >   {
> >   	u32 val;
> > @@ -1539,9 +1539,9 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> >   	val = I915_READ(sdvo_reg);
> >   	/* asserts want to know the pipe even if the port is disabled */
> > -	if (HAS_PCH_CPT(dev_priv))
> > +	if (HAS_PCH_CPT(i915))
> >   		*pipe = (val & SDVO_PIPE_SEL_MASK_CPT) >> SDVO_PIPE_SEL_SHIFT_CPT;
> > -	else if (IS_CHERRYVIEW(dev_priv))
> > +	else if (IS_CHERRYVIEW(i915))
> >   		*pipe = (val & SDVO_PIPE_SEL_MASK_CHV) >> SDVO_PIPE_SEL_SHIFT_CHV;
> >   	else
> >   		*pipe = (val & SDVO_PIPE_SEL_MASK) >> SDVO_PIPE_SEL_SHIFT;
> > @@ -1552,14 +1552,14 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> >   static bool intel_sdvo_get_hw_state(struct intel_encoder *encoder,
> >   				    enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> >   	u16 active_outputs = 0;
> >   	bool ret;
> >   	intel_sdvo_get_active_outputs(intel_sdvo, &active_outputs);
> > -	ret = intel_sdvo_port_enabled(dev_priv, intel_sdvo->sdvo_reg, pipe);
> > +	ret = intel_sdvo_port_enabled(i915, intel_sdvo->sdvo_reg, pipe);
> >   	return ret || active_outputs;
> >   }
> > @@ -1568,7 +1568,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder,
> >   				  struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> >   	struct intel_sdvo_dtd dtd;
> >   	int encoder_pixel_multiplier = 0;
> > @@ -1610,7 +1610,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder,
> >   	 * encoder->get_config we so already have a valid pixel multplier on all
> >   	 * other platfroms.
> >   	 */
> > -	if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) {
> > +	if (IS_I915G(i915) || IS_I915GM(i915)) {
> >   		pipe_config->pixel_multiplier =
> >   			((sdvox & SDVO_PORT_MULTIPLY_MASK)
> >   			 >> SDVO_PORT_MULTIPLY_SHIFT) + 1;
> > @@ -1693,7 +1693,7 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> >   			       const struct intel_crtc_state *old_crtc_state,
> >   			       const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> >   	u32 temp;
> > @@ -1716,13 +1716,13 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> >   	 * to transcoder A after disabling it to allow the
> >   	 * matching DP port to be enabled on transcoder A.
> >   	 */
> > -	if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) {
> > +	if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) {
> >   		/*
> >   		 * We get CPU/PCH FIFO underruns on the other pipe when
> >   		 * doing the workaround. Sweep them under the rug.
> >   		 */
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false);
> >   		temp &= ~SDVO_PIPE_SEL_MASK;
> >   		temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A);
> > @@ -1731,9 +1731,9 @@ static void intel_disable_sdvo(struct intel_encoder *encoder,
> >   		temp &= ~SDVO_ENABLE;
> >   		intel_sdvo_write_sdvox(intel_sdvo, temp);
> > -		intel_wait_for_vblank_if_active(dev_priv, PIPE_A);
> > -		intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > -		intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true);
> > +		intel_wait_for_vblank_if_active(i915, PIPE_A);
> > +		intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true);
> > +		intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true);
> >   	}
> >   }
> > @@ -1755,7 +1755,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder,
> >   			      const struct drm_connector_state *conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_sdvo *intel_sdvo = to_sdvo(encoder);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	u32 temp;
> > @@ -1768,7 +1768,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder,
> >   	intel_sdvo_write_sdvox(intel_sdvo, temp);
> >   	for (i = 0; i < 2; i++)
> > -		intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> > +		intel_wait_for_vblank(i915, intel_crtc->pipe);
> >   	success = intel_sdvo_get_trained_inputs(intel_sdvo, &input1, &input2);
> >   	/*
> > @@ -1865,17 +1865,17 @@ static bool intel_sdvo_get_capabilities(struct intel_sdvo *intel_sdvo, struct in
> >   static u16 intel_sdvo_get_hotplug_support(struct intel_sdvo *intel_sdvo)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(intel_sdvo->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(intel_sdvo->base.base.dev);
> >   	u16 hotplug;
> > -	if (!I915_HAS_HOTPLUG(dev_priv))
> > +	if (!I915_HAS_HOTPLUG(i915))
> >   		return 0;
> >   	/*
> >   	 * HW Erratum: SDVO Hotplug is broken on all i945G chips, there's noise
> >   	 * on the line.
> >   	 */
> > -	if (IS_I945G(dev_priv) || IS_I945GM(dev_priv))
> > +	if (IS_I945G(i915) || IS_I945GM(i915))
> >   		return 0;
> >   	if (!intel_sdvo_get_value(intel_sdvo, SDVO_CMD_GET_HOT_PLUG_SUPPORT,
> > @@ -1919,11 +1919,11 @@ intel_sdvo_get_edid(struct drm_connector *connector)
> >   static struct edid *
> >   intel_sdvo_get_analog_edid(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	return drm_get_edid(connector,
> > -			    intel_gmbus_get_adapter(dev_priv,
> > -						    dev_priv->vbt.crt_ddc_pin));
> > +			    intel_gmbus_get_adapter(i915,
> > +						    i915->vbt.crt_ddc_pin));
> >   }
> >   static enum drm_connector_status
> > @@ -2186,7 +2186,7 @@ static void intel_sdvo_get_tv_modes(struct drm_connector *connector)
> >   static void intel_sdvo_get_lvds_modes(struct drm_connector *connector)
> >   {
> >   	struct intel_sdvo *intel_sdvo = intel_attached_sdvo(connector);
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	struct drm_display_mode *newmode;
> >   	DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n",
> > @@ -2196,9 +2196,9 @@ static void intel_sdvo_get_lvds_modes(struct drm_connector *connector)
> >   	 * Fetch modes from VBT. For SDVO prefer the VBT mode since some
> >   	 * SDVO->LVDS transcoders can't cope with the EDID mode.
> >   	 */
> > -	if (dev_priv->vbt.sdvo_lvds_vbt_mode != NULL) {
> > +	if (i915->vbt.sdvo_lvds_vbt_mode != NULL) {
> >   		newmode = drm_mode_duplicate(connector->dev,
> > -					     dev_priv->vbt.sdvo_lvds_vbt_mode);
> > +					     i915->vbt.sdvo_lvds_vbt_mode);
> >   		if (newmode != NULL) {
> >   			/* Guarantee the mode is preferred */
> >   			newmode->type = (DRM_MODE_TYPE_PREFERRED |
> > @@ -2488,15 +2488,15 @@ intel_sdvo_guess_ddc_bus(struct intel_sdvo *sdvo)
> >    * outputs, then LVDS outputs.
> >    */
> >   static void
> > -intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv,
> > +intel_sdvo_select_ddc_bus(struct drm_i915_private *i915,
> >   			  struct intel_sdvo *sdvo)
> >   {
> >   	struct sdvo_device_mapping *mapping;
> >   	if (sdvo->port == PORT_B)
> > -		mapping = &dev_priv->vbt.sdvo_mappings[0];
> > +		mapping = &i915->vbt.sdvo_mappings[0];
> >   	else
> > -		mapping = &dev_priv->vbt.sdvo_mappings[1];
> > +		mapping = &i915->vbt.sdvo_mappings[1];
> >   	if (mapping->initialized)
> >   		sdvo->ddc_bus = 1 << ((mapping->ddc_pin & 0xf0) >> 4);
> > @@ -2505,24 +2505,24 @@ intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv,
> >   }
> >   static void
> > -intel_sdvo_select_i2c_bus(struct drm_i915_private *dev_priv,
> > +intel_sdvo_select_i2c_bus(struct drm_i915_private *i915,
> >   			  struct intel_sdvo *sdvo)
> >   {
> >   	struct sdvo_device_mapping *mapping;
> >   	u8 pin;
> >   	if (sdvo->port == PORT_B)
> > -		mapping = &dev_priv->vbt.sdvo_mappings[0];
> > +		mapping = &i915->vbt.sdvo_mappings[0];
> >   	else
> > -		mapping = &dev_priv->vbt.sdvo_mappings[1];
> > +		mapping = &i915->vbt.sdvo_mappings[1];
> >   	if (mapping->initialized &&
> > -	    intel_gmbus_is_valid_pin(dev_priv, mapping->i2c_pin))
> > +	    intel_gmbus_is_valid_pin(i915, mapping->i2c_pin))
> >   		pin = mapping->i2c_pin;
> >   	else
> >   		pin = GMBUS_PIN_DPB;
> > -	sdvo->i2c = intel_gmbus_get_adapter(dev_priv, pin);
> > +	sdvo->i2c = intel_gmbus_get_adapter(i915, pin);
> >   	/*
> >   	 * With gmbus we should be able to drive sdvo i2c at 2MHz, but somehow
> > @@ -2546,17 +2546,17 @@ intel_sdvo_is_hdmi_connector(struct intel_sdvo *intel_sdvo, int device)
> >   }
> >   static u8
> > -intel_sdvo_get_slave_addr(struct drm_i915_private *dev_priv,
> > +intel_sdvo_get_slave_addr(struct drm_i915_private *i915,
> >   			  struct intel_sdvo *sdvo)
> >   {
> >   	struct sdvo_device_mapping *my_mapping, *other_mapping;
> >   	if (sdvo->port == PORT_B) {
> > -		my_mapping = &dev_priv->vbt.sdvo_mappings[0];
> > -		other_mapping = &dev_priv->vbt.sdvo_mappings[1];
> > +		my_mapping = &i915->vbt.sdvo_mappings[0];
> > +		other_mapping = &i915->vbt.sdvo_mappings[1];
> >   	} else {
> > -		my_mapping = &dev_priv->vbt.sdvo_mappings[1];
> > -		other_mapping = &dev_priv->vbt.sdvo_mappings[0];
> > +		my_mapping = &i915->vbt.sdvo_mappings[1];
> > +		other_mapping = &i915->vbt.sdvo_mappings[0];
> >   	}
> >   	/* If the BIOS described our SDVO device, take advantage of it. */
> > @@ -2616,10 +2616,10 @@ static void
> >   intel_sdvo_add_hdmi_properties(struct intel_sdvo *intel_sdvo,
> >   			       struct intel_sdvo_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.base.dev);
> >   	intel_attach_force_audio_property(&connector->base.base);
> > -	if (INTEL_GEN(dev_priv) >= 4 && IS_MOBILE(dev_priv)) {
> > +	if (INTEL_GEN(i915) >= 4 && IS_MOBILE(i915)) {
> >   		intel_attach_broadcast_rgb_property(&connector->base.base);
> >   	}
> >   	intel_attach_aspect_ratio_property(&connector->base.base);
> > @@ -3180,9 +3180,9 @@ static const struct i2c_lock_operations proxy_lock_ops = {
> >   static bool
> >   intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo,
> > -			  struct drm_i915_private *dev_priv)
> > +			  struct drm_i915_private *i915)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	sdvo->ddc.owner = THIS_MODULE;
> >   	sdvo->ddc.class = I2C_CLASS_DDC;
> > @@ -3195,23 +3195,23 @@ intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo,
> >   	return i2c_add_adapter(&sdvo->ddc) == 0;
> >   }
> > -static void assert_sdvo_port_valid(const struct drm_i915_private *dev_priv,
> > +static void assert_sdvo_port_valid(const struct drm_i915_private *i915,
> >   				   enum port port)
> >   {
> > -	if (HAS_PCH_SPLIT(dev_priv))
> > +	if (HAS_PCH_SPLIT(i915))
> >   		WARN_ON(port != PORT_B);
> >   	else
> >   		WARN_ON(port != PORT_B && port != PORT_C);
> >   }
> > -bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> > +bool intel_sdvo_init(struct drm_i915_private *i915,
> >   		     i915_reg_t sdvo_reg, enum port port)
> >   {
> >   	struct intel_encoder *intel_encoder;
> >   	struct intel_sdvo *intel_sdvo;
> >   	int i;
> > -	assert_sdvo_port_valid(dev_priv, port);
> > +	assert_sdvo_port_valid(i915, port);
> >   	intel_sdvo = kzalloc(sizeof(*intel_sdvo), GFP_KERNEL);
> >   	if (!intel_sdvo)
> > @@ -3220,9 +3220,9 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> >   	intel_sdvo->sdvo_reg = sdvo_reg;
> >   	intel_sdvo->port = port;
> >   	intel_sdvo->slave_addr =
> > -		intel_sdvo_get_slave_addr(dev_priv, intel_sdvo) >> 1;
> > -	intel_sdvo_select_i2c_bus(dev_priv, intel_sdvo);
> > -	if (!intel_sdvo_init_ddc_proxy(intel_sdvo, dev_priv))
> > +		intel_sdvo_get_slave_addr(i915, intel_sdvo) >> 1;
> > +	intel_sdvo_select_i2c_bus(i915, intel_sdvo);
> > +	if (!intel_sdvo_init_ddc_proxy(intel_sdvo, i915))
> >   		goto err_i2c_bus;
> >   	/* encoder type will be decided later */
> > @@ -3230,7 +3230,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> >   	intel_encoder->type = INTEL_OUTPUT_SDVO;
> >   	intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER;
> >   	intel_encoder->port = port;
> > -	drm_encoder_init(&dev_priv->drm, &intel_encoder->base,
> > +	drm_encoder_init(&i915->drm, &intel_encoder->base,
> >   			 &intel_sdvo_enc_funcs, 0,
> >   			 "SDVO %c", port_name(port));
> > @@ -3246,7 +3246,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> >   	}
> >   	intel_encoder->compute_config = intel_sdvo_compute_config;
> > -	if (HAS_PCH_SPLIT(dev_priv)) {
> > +	if (HAS_PCH_SPLIT(i915)) {
> >   		intel_encoder->disable = pch_disable_sdvo;
> >   		intel_encoder->post_disable = pch_post_disable_sdvo;
> >   	} else {
> > @@ -3290,7 +3290,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> >   	 */
> >   	intel_sdvo->base.cloneable = 0;
> > -	intel_sdvo_select_ddc_bus(dev_priv, intel_sdvo);
> > +	intel_sdvo_select_ddc_bus(i915, intel_sdvo);
> >   	/* Set the input timing to the screen. Assume always input 0. */
> >   	if (!intel_sdvo_set_target_input(intel_sdvo))
> > diff --git a/drivers/gpu/drm/i915/intel_sdvo.h b/drivers/gpu/drm/i915/intel_sdvo.h
> > index c9e05bcdd141..1ec6766c5ae1 100644
> > --- a/drivers/gpu/drm/i915/intel_sdvo.h
> > +++ b/drivers/gpu/drm/i915/intel_sdvo.h
> > @@ -15,9 +15,9 @@
> >   struct drm_i915_private;
> >   enum pipe;
> > -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv,
> > +bool intel_sdvo_port_enabled(struct drm_i915_private *i915,
> >   			     i915_reg_t sdvo_reg, enum pipe *pipe);
> > -bool intel_sdvo_init(struct drm_i915_private *dev_priv,
> > +bool intel_sdvo_init(struct drm_i915_private *i915,
> >   		     i915_reg_t reg, enum port port);
> >   #endif /* __INTEL_SDVO_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_sprite.c b/drivers/gpu/drm/i915/intel_sprite.c
> > index c180815faabd..73871e5e8e19 100644
> > --- a/drivers/gpu/drm/i915/intel_sprite.c
> > +++ b/drivers/gpu/drm/i915/intel_sprite.c
> > @@ -94,12 +94,12 @@ int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode,
> >   void intel_pipe_update_start(const struct intel_crtc_state *new_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	const struct drm_display_mode *adjusted_mode = &new_crtc_state->base.adjusted_mode;
> >   	long timeout = msecs_to_jiffies_timeout(1);
> >   	int scanline, min, max, vblank_start;
> >   	wait_queue_head_t *wq = drm_crtc_vblank_waitqueue(&crtc->base);
> > -	bool need_vlv_dsi_wa = (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	bool need_vlv_dsi_wa = (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   		intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI);
> >   	DEFINE_WAIT(wait);
> >   	u32 psr_status;
> > @@ -207,7 +207,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state)
> >   	int scanline_end = intel_get_crtc_scanline(crtc);
> >   	u32 end_vbl_count = intel_crtc_get_vblank_counter(crtc);
> >   	ktime_t end_vbl_time = ktime_get();
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	trace_i915_pipe_update_end(crtc, end_vbl_count, scanline_end);
> > @@ -227,7 +227,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state)
> >   	local_irq_enable();
> > -	if (intel_vgpu_active(dev_priv))
> > +	if (intel_vgpu_active(i915))
> >   		return;
> >   	if (crtc->debug.start_vbl_count &&
> > @@ -353,7 +353,7 @@ skl_program_scaler(struct intel_plane *plane,
> >   		   const struct intel_crtc_state *crtc_state,
> >   		   const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	int scaler_id = plane_state->scaler_id;
> >   	const struct intel_scaler *scaler =
> > @@ -375,7 +375,7 @@ skl_program_scaler(struct intel_plane *plane,
> >   	/* TODO: handle sub-pixel coordinates */
> >   	if (is_planar_yuv_format(plane_state->base.fb->format->format) &&
> > -	    !icl_is_hdr_plane(dev_priv, plane->id)) {
> > +	    !icl_is_hdr_plane(i915, plane->id)) {
> >   		y_hphase = skl_scaler_calc_phase(1, hscale, false);
> >   		y_vphase = skl_scaler_calc_phase(1, vscale, false);
> > @@ -415,7 +415,7 @@ icl_program_input_csc(struct intel_plane *plane,
> >   		      const struct intel_crtc_state *crtc_state,
> >   		      const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	enum plane_id plane_id = plane->id;
> > @@ -507,7 +507,7 @@ skl_program_plane(struct intel_plane *plane,
> >   		  const struct intel_plane_state *plane_state,
> >   		  int color_plane, bool slave, u32 plane_ctl)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum plane_id plane_id = plane->id;
> >   	enum pipe pipe = plane->pipe;
> >   	const struct drm_intel_sprite_colorkey *key = &plane_state->ckey;
> > @@ -529,7 +529,7 @@ skl_program_plane(struct intel_plane *plane,
> >   	plane_ctl |= skl_plane_ctl_crtc(crtc_state);
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		plane_color_ctl = plane_state->color_ctl |
> >   			glk_plane_color_ctl_crtc(crtc_state);
> > @@ -549,7 +549,7 @@ skl_program_plane(struct intel_plane *plane,
> >   		crtc_y = 0;
> >   	}
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(PLANE_STRIDE(pipe, plane_id), stride);
> >   	I915_WRITE_FW(PLANE_POS(pipe, plane_id), (crtc_y << 16) | crtc_x);
> > @@ -557,7 +557,7 @@ skl_program_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(PLANE_AUX_DIST(pipe, plane_id),
> >   		      (plane_state->color_plane[1].offset - surf_addr) | aux_stride);
> > -	if (icl_is_hdr_plane(dev_priv, plane_id)) {
> > +	if (icl_is_hdr_plane(i915, plane_id)) {
> >   		u32 cus_ctl = 0;
> >   		if (linked) {
> > @@ -578,10 +578,10 @@ skl_program_plane(struct intel_plane *plane,
> >   		I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), cus_ctl);
> >   	}
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		I915_WRITE_FW(PLANE_COLOR_CTL(pipe, plane_id), plane_color_ctl);
> > -	if (fb->format->is_yuv && icl_is_hdr_plane(dev_priv, plane_id))
> > +	if (fb->format->is_yuv && icl_is_hdr_plane(i915, plane_id))
> >   		icl_program_input_csc(plane, crtc_state, plane_state);
> >   	skl_write_plane_wm(plane, crtc_state);
> > @@ -592,7 +592,7 @@ skl_program_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(PLANE_OFFSET(pipe, plane_id), (y << 16) | x);
> > -	if (INTEL_GEN(dev_priv) < 11)
> > +	if (INTEL_GEN(i915) < 11)
> >   		I915_WRITE_FW(PLANE_AUX_OFFSET(pipe, plane_id),
> >   			      (plane_state->color_plane[1].y << 16) |
> >   			      plane_state->color_plane[1].x);
> > @@ -609,7 +609,7 @@ skl_program_plane(struct intel_plane *plane,
> >   	if (!slave && plane_state->scaler_id >= 0)
> >   		skl_program_scaler(plane, crtc_state, plane_state);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void
> > @@ -641,14 +641,14 @@ static void
> >   skl_disable_plane(struct intel_plane *plane,
> >   		  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum plane_id plane_id = plane->id;
> >   	enum pipe pipe = plane->pipe;
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> > -	if (icl_is_hdr_plane(dev_priv, plane_id))
> > +	if (icl_is_hdr_plane(i915, plane_id))
> >   		I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), 0);
> >   	skl_write_plane_wm(plane, crtc_state);
> > @@ -656,21 +656,21 @@ skl_disable_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(PLANE_CTL(pipe, plane_id), 0);
> >   	I915_WRITE_FW(PLANE_SURF(pipe, plane_id), 0);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static bool
> >   skl_plane_get_hw_state(struct intel_plane *plane,
> >   		       enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	enum plane_id plane_id = plane->id;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -678,7 +678,7 @@ skl_plane_get_hw_state(struct intel_plane *plane,
> >   	*pipe = plane->pipe;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -687,7 +687,7 @@ static void
> >   chv_update_csc(const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	enum plane_id plane_id = plane->id;
> >   	/*
> > @@ -746,7 +746,7 @@ static void
> >   vlv_update_clrc(const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	enum pipe pipe = plane->pipe;
> >   	enum plane_id plane_id = plane->id;
> > @@ -863,7 +863,7 @@ vlv_update_plane(struct intel_plane *plane,
> >   		 const struct intel_crtc_state *crtc_state,
> >   		 const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	enum plane_id plane_id = plane->id;
> >   	u32 sprsurf_offset = plane_state->color_plane[0].offset;
> > @@ -886,7 +886,7 @@ vlv_update_plane(struct intel_plane *plane,
> >   	linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(SPSTRIDE(pipe, plane_id),
> >   		      plane_state->color_plane[0].stride);
> > @@ -894,7 +894,7 @@ vlv_update_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(SPSIZE(pipe, plane_id), (crtc_h << 16) | crtc_w);
> >   	I915_WRITE_FW(SPCONSTALPHA(pipe, plane_id), 0);
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B)
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_B)
> >   		chv_update_csc(plane_state);
> >   	if (key->flags) {
> > @@ -917,38 +917,38 @@ vlv_update_plane(struct intel_plane *plane,
> >   	vlv_update_clrc(plane_state);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void
> >   vlv_disable_plane(struct intel_plane *plane,
> >   		  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	enum plane_id plane_id = plane->id;
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(SPCNTR(pipe, plane_id), 0);
> >   	I915_WRITE_FW(SPSURF(pipe, plane_id), 0);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static bool
> >   vlv_plane_get_hw_state(struct intel_plane *plane,
> >   		       enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	enum plane_id plane_id = plane->id;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -956,7 +956,7 @@ vlv_plane_get_hw_state(struct intel_plane *plane,
> >   	*pipe = plane->pipe;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -977,7 +977,7 @@ static u32 ivb_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state,
> >   			  const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> > @@ -986,7 +986,7 @@ static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state,
> >   	sprctl = SPRITE_ENABLE;
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		sprctl |= SPRITE_TRICKLE_FEED_DISABLE;
> >   	switch (fb->format->format) {
> > @@ -1038,7 +1038,7 @@ ivb_update_plane(struct intel_plane *plane,
> >   		 const struct intel_crtc_state *crtc_state,
> >   		 const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	u32 sprsurf_offset = plane_state->color_plane[0].offset;
> >   	u32 linear_offset;
> > @@ -1067,12 +1067,12 @@ ivb_update_plane(struct intel_plane *plane,
> >   	linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(SPRSTRIDE(pipe), plane_state->color_plane[0].stride);
> >   	I915_WRITE_FW(SPRPOS(pipe), (crtc_y << 16) | crtc_x);
> >   	I915_WRITE_FW(SPRSIZE(pipe), (crtc_h << 16) | crtc_w);
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		I915_WRITE_FW(SPRSCALE(pipe), sprscale);
> >   	if (key->flags) {
> > @@ -1083,7 +1083,7 @@ ivb_update_plane(struct intel_plane *plane,
> >   	/* HSW consolidates SPRTILEOFF and SPRLINOFF into a single SPROFFSET
> >   	 * register */
> > -	if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
> > +	if (IS_HASWELL(i915) || IS_BROADWELL(i915)) {
> >   		I915_WRITE_FW(SPROFFSET(pipe), (y << 16) | x);
> >   	} else {
> >   		I915_WRITE_FW(SPRLINOFF(pipe), linear_offset);
> > @@ -1099,39 +1099,39 @@ ivb_update_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(SPRSURF(pipe),
> >   		      intel_plane_ggtt_offset(plane_state) + sprsurf_offset);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void
> >   ivb_disable_plane(struct intel_plane *plane,
> >   		  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(SPRCTL(pipe), 0);
> >   	/* Disable the scaler */
> > -	if (IS_IVYBRIDGE(dev_priv))
> > +	if (IS_IVYBRIDGE(i915))
> >   		I915_WRITE_FW(SPRSCALE(pipe), 0);
> >   	I915_WRITE_FW(SPRSURF(pipe), 0);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static bool
> >   ivb_plane_get_hw_state(struct intel_plane *plane,
> >   		       enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1139,7 +1139,7 @@ ivb_plane_get_hw_state(struct intel_plane *plane,
> >   	*pipe = plane->pipe;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -1168,7 +1168,7 @@ static u32 g4x_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state)
> >   static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state,
> >   			  const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> > @@ -1177,7 +1177,7 @@ static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state,
> >   	dvscntr = DVS_ENABLE;
> > -	if (IS_GEN(dev_priv, 6))
> > +	if (IS_GEN(i915, 6))
> >   		dvscntr |= DVS_TRICKLE_FEED_DISABLE;
> >   	switch (fb->format->format) {
> > @@ -1229,7 +1229,7 @@ g4x_update_plane(struct intel_plane *plane,
> >   		 const struct intel_crtc_state *crtc_state,
> >   		 const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	u32 dvssurf_offset = plane_state->color_plane[0].offset;
> >   	u32 linear_offset;
> > @@ -1258,7 +1258,7 @@ g4x_update_plane(struct intel_plane *plane,
> >   	linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0);
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(DVSSTRIDE(pipe), plane_state->color_plane[0].stride);
> >   	I915_WRITE_FW(DVSPOS(pipe), (crtc_y << 16) | crtc_x);
> > @@ -1283,38 +1283,38 @@ g4x_update_plane(struct intel_plane *plane,
> >   	I915_WRITE_FW(DVSSURF(pipe),
> >   		      intel_plane_ggtt_offset(plane_state) + dvssurf_offset);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static void
> >   g4x_disable_plane(struct intel_plane *plane,
> >   		  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum pipe pipe = plane->pipe;
> >   	unsigned long irqflags;
> > -	spin_lock_irqsave(&dev_priv->uncore.lock, irqflags);
> > +	spin_lock_irqsave(&i915->uncore.lock, irqflags);
> >   	I915_WRITE_FW(DVSCNTR(pipe), 0);
> >   	/* Disable the scaler */
> >   	I915_WRITE_FW(DVSSCALE(pipe), 0);
> >   	I915_WRITE_FW(DVSSURF(pipe), 0);
> > -	spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags);
> > +	spin_unlock_irqrestore(&i915->uncore.lock, irqflags);
> >   }
> >   static bool
> >   g4x_plane_get_hw_state(struct intel_plane *plane,
> >   		       enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	enum intel_display_power_domain power_domain;
> >   	intel_wakeref_t wakeref;
> >   	bool ret;
> >   	power_domain = POWER_DOMAIN_PIPE(plane->pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1322,7 +1322,7 @@ g4x_plane_get_hw_state(struct intel_plane *plane,
> >   	*pipe = plane->pipe;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return ret;
> >   }
> > @@ -1406,16 +1406,16 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> >   		 struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	int min_scale = DRM_PLANE_HELPER_NO_SCALING;
> >   	int max_scale = DRM_PLANE_HELPER_NO_SCALING;
> >   	int ret;
> >   	if (intel_fb_scalable(plane_state->base.fb)) {
> > -		if (INTEL_GEN(dev_priv) < 7) {
> > +		if (INTEL_GEN(i915) < 7) {
> >   			min_scale = 1;
> >   			max_scale = 16 << 16;
> > -		} else if (IS_IVYBRIDGE(dev_priv)) {
> > +		} else if (IS_IVYBRIDGE(i915)) {
> >   			min_scale = 1;
> >   			max_scale = 2 << 16;
> >   		}
> > @@ -1443,7 +1443,7 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> >   	if (ret)
> >   		return ret;
> > -	if (INTEL_GEN(dev_priv) >= 7)
> > +	if (INTEL_GEN(i915) >= 7)
> >   		plane_state->ctl = ivb_sprite_ctl(crtc_state, plane_state);
> >   	else
> >   		plane_state->ctl = g4x_sprite_ctl(crtc_state, plane_state);
> > @@ -1454,11 +1454,11 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state,
> >   int chv_plane_check_rotation(const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	unsigned int rotation = plane_state->base.rotation;
> >   	/* CHV ignores the mirror bit when the rotate bit is set :( */
> > -	if (IS_CHERRYVIEW(dev_priv) &&
> > +	if (IS_CHERRYVIEW(i915) &&
> >   	    rotation & DRM_MODE_ROTATE_180 &&
> >   	    rotation & DRM_MODE_REFLECT_X) {
> >   		DRM_DEBUG_KMS("Cannot rotate and reflect at the same time\n");
> > @@ -1506,7 +1506,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> >   			      const struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	unsigned int rotation = plane_state->base.rotation;
> >   	struct drm_format_name_buf format_name;
> > @@ -1540,7 +1540,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> >   		 */
> >   		switch (fb->format->format) {
> >   		case DRM_FORMAT_RGB565:
> > -			if (INTEL_GEN(dev_priv) >= 11)
> > +			if (INTEL_GEN(i915) >= 11)
> >   				break;
> >   			/* fall through */
> >   		case DRM_FORMAT_C8:
> > @@ -1579,7 +1579,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state,
> >   static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_state,
> >   					   const struct intel_plane_state *plane_state)
> >   {
> > -	struct drm_i915_private *dev_priv =
> > +	struct drm_i915_private *i915 =
> >   		to_i915(plane_state->base.plane->dev);
> >   	int crtc_x = plane_state->base.dst.x1;
> >   	int crtc_w = drm_rect_width(&plane_state->base.dst);
> > @@ -1594,7 +1594,7 @@ static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_s
> >   	 * than the cursor ending less than 4 pixels from the left edge of the
> >   	 * screen may cause FIFO underflow and display corruption.
> >   	 */
> > -	if ((IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) &&
> > +	if ((IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) &&
> >   	    (crtc_x + crtc_w < 4 || crtc_x > pipe_src_w - 4)) {
> >   		DRM_DEBUG_KMS("requested plane X %s position %d invalid (valid range %d-%d)\n",
> >   			      crtc_x + crtc_w < 4 ? "end" : "start",
> > @@ -1627,7 +1627,7 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state,
> >   			   struct intel_plane_state *plane_state)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	const struct drm_framebuffer *fb = plane_state->base.fb;
> >   	int min_scale = DRM_PLANE_HELPER_NO_SCALING;
> >   	int max_scale = DRM_PLANE_HELPER_NO_SCALING;
> > @@ -1675,23 +1675,23 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state,
> >   	plane_state->ctl = skl_plane_ctl(crtc_state, plane_state);
> > -	if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))
> > +	if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))
> >   		plane_state->color_ctl = glk_plane_color_ctl(crtc_state,
> >   							     plane_state);
> >   	return 0;
> >   }
> > -static bool has_dst_key_in_primary_plane(struct drm_i915_private *dev_priv)
> > +static bool has_dst_key_in_primary_plane(struct drm_i915_private *i915)
> >   {
> > -	return INTEL_GEN(dev_priv) >= 9;
> > +	return INTEL_GEN(i915) >= 9;
> >   }
> >   static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> >   				 const struct drm_intel_sprite_colorkey *set)
> >   {
> >   	struct intel_plane *plane = to_intel_plane(plane_state->base.plane);
> > -	struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(plane->base.dev);
> >   	struct drm_intel_sprite_colorkey *key = &plane_state->ckey;
> >   	*key = *set;
> > @@ -1708,7 +1708,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> >   	 * On SKL+ we want dst key enabled on
> >   	 * the primary and not on the sprite.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_PRIMARY &&
> > +	if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_PRIMARY &&
> >   	    set->flags & I915_SET_COLORKEY_DESTINATION)
> >   		key->flags = 0;
> >   }
> > @@ -1716,7 +1716,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state,
> >   int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> >   				    struct drm_file *file_priv)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_intel_sprite_colorkey *set = data;
> >   	struct drm_plane *plane;
> >   	struct drm_plane_state *plane_state;
> > @@ -1734,7 +1734,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> >   	if ((set->flags & (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE)) == (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE))
> >   		return -EINVAL;
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   	    set->flags & I915_SET_COLORKEY_DESTINATION)
> >   		return -EINVAL;
> > @@ -1747,7 +1747,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> >   	 * Also multiple planes can't do destination keying on the same
> >   	 * pipe simultaneously.
> >   	 */
> > -	if (INTEL_GEN(dev_priv) >= 9 &&
> > +	if (INTEL_GEN(i915) >= 9 &&
> >   	    to_intel_plane(plane)->id >= PLANE_SPRITE1 &&
> >   	    set->flags & I915_SET_COLORKEY_DESTINATION)
> >   		return -EINVAL;
> > @@ -1771,9 +1771,9 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> >   		 * On some platforms we have to configure
> >   		 * the dst colorkey on the primary plane.
> >   		 */
> > -		if (!ret && has_dst_key_in_primary_plane(dev_priv)) {
> > +		if (!ret && has_dst_key_in_primary_plane(i915)) {
> >   			struct intel_crtc *crtc =
> > -				intel_get_crtc_for_pipe(dev_priv,
> > +				intel_get_crtc_for_pipe(i915,
> >   							to_intel_plane(plane)->pipe);
> >   			plane_state = drm_atomic_get_plane_state(state,
> > @@ -2197,26 +2197,26 @@ static const struct drm_plane_funcs skl_plane_funcs = {
> >   	.format_mod_supported = skl_plane_format_mod_supported,
> >   };
> > -static bool skl_plane_has_fbc(struct drm_i915_private *dev_priv,
> > +static bool skl_plane_has_fbc(struct drm_i915_private *i915,
> >   			      enum pipe pipe, enum plane_id plane_id)
> >   {
> > -	if (!HAS_FBC(dev_priv))
> > +	if (!HAS_FBC(i915))
> >   		return false;
> >   	return pipe == PIPE_A && plane_id == PLANE_PRIMARY;
> >   }
> > -static bool skl_plane_has_planar(struct drm_i915_private *dev_priv,
> > +static bool skl_plane_has_planar(struct drm_i915_private *i915,
> >   				 enum pipe pipe, enum plane_id plane_id)
> >   {
> > -	if (INTEL_GEN(dev_priv) >= 11)
> > +	if (INTEL_GEN(i915) >= 11)
> >   		return plane_id <= PLANE_SPRITE3;
> >   	/* Display WA #0870: skl, bxt */
> > -	if (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv))
> > +	if (IS_SKYLAKE(i915) || IS_BROXTON(i915))
> >   		return false;
> > -	if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv) && pipe == PIPE_C)
> > +	if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915) && pipe == PIPE_C)
> >   		return false;
> >   	if (plane_id != PLANE_PRIMARY && plane_id != PLANE_SPRITE0)
> > @@ -2225,16 +2225,16 @@ static bool skl_plane_has_planar(struct drm_i915_private *dev_priv,
> >   	return true;
> >   }
> > -static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv,
> > +static bool skl_plane_has_ccs(struct drm_i915_private *i915,
> >   			      enum pipe pipe, enum plane_id plane_id)
> >   {
> >   	if (plane_id == PLANE_CURSOR)
> >   		return false;
> > -	if (INTEL_GEN(dev_priv) >= 10)
> > +	if (INTEL_GEN(i915) >= 10)
> >   		return true;
> > -	if (IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEMINILAKE(i915))
> >   		return pipe != PIPE_C;
> >   	return pipe != PIPE_C &&
> > @@ -2243,7 +2243,7 @@ static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv,
> >   }
> >   struct intel_plane *
> > -skl_universal_plane_create(struct drm_i915_private *dev_priv,
> > +skl_universal_plane_create(struct drm_i915_private *i915,
> >   			   enum pipe pipe, enum plane_id plane_id)
> >   {
> >   	struct intel_plane *plane;
> > @@ -2263,9 +2263,9 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   	plane->id = plane_id;
> >   	plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane_id);
> > -	plane->has_fbc = skl_plane_has_fbc(dev_priv, pipe, plane_id);
> > +	plane->has_fbc = skl_plane_has_fbc(i915, pipe, plane_id);
> >   	if (plane->has_fbc) {
> > -		struct intel_fbc *fbc = &dev_priv->fbc;
> > +		struct intel_fbc *fbc = &i915->fbc;
> >   		fbc->possible_framebuffer_bits |= plane->frontbuffer_bit;
> >   	}
> > @@ -2278,24 +2278,24 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   	if (icl_is_nv12_y_plane(plane_id))
> >   		plane->update_slave = icl_update_slave;
> > -	if (skl_plane_has_planar(dev_priv, pipe, plane_id)) {
> > -		if (icl_is_hdr_plane(dev_priv, plane_id)) {
> > +	if (skl_plane_has_planar(i915, pipe, plane_id)) {
> > +		if (icl_is_hdr_plane(i915, plane_id)) {
> >   			formats = icl_hdr_planar_formats;
> >   			num_formats = ARRAY_SIZE(icl_hdr_planar_formats);
> > -		} else if (INTEL_GEN(dev_priv) >= 11) {
> > +		} else if (INTEL_GEN(i915) >= 11) {
> >   			formats = icl_planar_formats;
> >   			num_formats = ARRAY_SIZE(icl_planar_formats);
> > -		} else if (INTEL_GEN(dev_priv) == 10 || IS_GEMINILAKE(dev_priv)) {
> > +		} else if (INTEL_GEN(i915) == 10 || IS_GEMINILAKE(i915)) {
> >   			formats = glk_planar_formats;
> >   			num_formats = ARRAY_SIZE(glk_planar_formats);
> >   		} else {
> >   			formats = skl_planar_formats;
> >   			num_formats = ARRAY_SIZE(skl_planar_formats);
> >   		}
> > -	} else if (icl_is_hdr_plane(dev_priv, plane_id)) {
> > +	} else if (icl_is_hdr_plane(i915, plane_id)) {
> >   		formats = icl_hdr_plane_formats;
> >   		num_formats = ARRAY_SIZE(icl_hdr_plane_formats);
> > -	} else if (INTEL_GEN(dev_priv) >= 11) {
> > +	} else if (INTEL_GEN(i915) >= 11) {
> >   		formats = icl_plane_formats;
> >   		num_formats = ARRAY_SIZE(icl_plane_formats);
> >   	} else {
> > @@ -2303,7 +2303,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   		num_formats = ARRAY_SIZE(skl_plane_formats);
> >   	}
> > -	plane->has_ccs = skl_plane_has_ccs(dev_priv, pipe, plane_id);
> > +	plane->has_ccs = skl_plane_has_ccs(i915, pipe, plane_id);
> >   	if (plane->has_ccs)
> >   		modifiers = skl_plane_format_modifiers_ccs;
> >   	else
> > @@ -2316,7 +2316,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   	possible_crtcs = BIT(pipe);
> > -	ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> > +	ret = drm_universal_plane_init(&i915->drm, &plane->base,
> >   				       possible_crtcs, &skl_plane_funcs,
> >   				       formats, num_formats, modifiers,
> >   				       plane_type,
> > @@ -2329,7 +2329,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   		DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_90 |
> >   		DRM_MODE_ROTATE_180 | DRM_MODE_ROTATE_270;
> > -	if (INTEL_GEN(dev_priv) >= 10)
> > +	if (INTEL_GEN(i915) >= 10)
> >   		supported_rotations |= DRM_MODE_REFLECT_X;
> >   	drm_plane_create_rotation_property(&plane->base,
> > @@ -2361,7 +2361,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv,
> >   }
> >   struct intel_plane *
> > -intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> > +intel_sprite_plane_create(struct drm_i915_private *i915,
> >   			  enum pipe pipe, int sprite)
> >   {
> >   	struct intel_plane *plane;
> > @@ -2373,15 +2373,15 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> >   	int num_formats;
> >   	int ret;
> > -	if (INTEL_GEN(dev_priv) >= 9)
> > -		return skl_universal_plane_create(dev_priv, pipe,
> > +	if (INTEL_GEN(i915) >= 9)
> > +		return skl_universal_plane_create(i915, pipe,
> >   						  PLANE_SPRITE0 + sprite);
> >   	plane = intel_plane_alloc();
> >   	if (IS_ERR(plane))
> >   		return plane;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		plane->max_stride = i9xx_plane_max_stride;
> >   		plane->update_plane = vlv_update_plane;
> >   		plane->disable_plane = vlv_disable_plane;
> > @@ -2393,7 +2393,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> >   		modifiers = i9xx_plane_format_modifiers;
> >   		plane_funcs = &vlv_sprite_funcs;
> > -	} else if (INTEL_GEN(dev_priv) >= 7) {
> > +	} else if (INTEL_GEN(i915) >= 7) {
> >   		plane->max_stride = g4x_sprite_max_stride;
> >   		plane->update_plane = ivb_update_plane;
> >   		plane->disable_plane = ivb_disable_plane;
> > @@ -2413,7 +2413,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> >   		plane->check_plane = g4x_sprite_check;
> >   		modifiers = i9xx_plane_format_modifiers;
> > -		if (IS_GEN(dev_priv, 6)) {
> > +		if (IS_GEN(i915, 6)) {
> >   			formats = snb_plane_formats;
> >   			num_formats = ARRAY_SIZE(snb_plane_formats);
> > @@ -2426,7 +2426,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> >   		}
> >   	}
> > -	if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) {
> > +	if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) {
> >   		supported_rotations =
> >   			DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 |
> >   			DRM_MODE_REFLECT_X;
> > @@ -2441,7 +2441,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> >   	possible_crtcs = BIT(pipe);
> > -	ret = drm_universal_plane_init(&dev_priv->drm, &plane->base,
> > +	ret = drm_universal_plane_init(&i915->drm, &plane->base,
> >   				       possible_crtcs, plane_funcs,
> >   				       formats, num_formats, modifiers,
> >   				       DRM_PLANE_TYPE_OVERLAY,
> > diff --git a/drivers/gpu/drm/i915/intel_sprite.h b/drivers/gpu/drm/i915/intel_sprite.h
> > index 500f6bffb139..401295e05116 100644
> > --- a/drivers/gpu/drm/i915/intel_sprite.h
> > +++ b/drivers/gpu/drm/i915/intel_sprite.h
> > @@ -21,7 +21,7 @@ struct intel_plane_state;
> >   bool is_planar_yuv_format(u32 pixelformat);
> >   int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode,
> >   			     int usecs);
> > -struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *dev_priv,
> > +struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *i915,
> >   					      enum pipe pipe, int plane);
> >   int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data,
> >   				    struct drm_file *file_priv);
> > @@ -31,7 +31,7 @@ int intel_plane_check_stride(const struct intel_plane_state *plane_state);
> >   int intel_plane_check_src_coordinates(struct intel_plane_state *plane_state);
> >   int chv_plane_check_rotation(const struct intel_plane_state *plane_state);
> >   struct intel_plane *
> > -skl_universal_plane_create(struct drm_i915_private *dev_priv,
> > +skl_universal_plane_create(struct drm_i915_private *i915,
> >   			   enum pipe pipe, enum plane_id plane_id);
> >   static inline bool icl_is_nv12_y_plane(enum plane_id id)
> > @@ -49,10 +49,10 @@ static inline u8 icl_hdr_plane_mask(void)
> >   		BIT(PLANE_SPRITE0) | BIT(PLANE_SPRITE1);
> >   }
> > -static inline bool icl_is_hdr_plane(struct drm_i915_private *dev_priv,
> > +static inline bool icl_is_hdr_plane(struct drm_i915_private *i915,
> >   				    enum plane_id plane_id)
> >   {
> > -	return INTEL_GEN(dev_priv) >= 11 &&
> > +	return INTEL_GEN(i915) >= 11 &&
> >   		icl_hdr_plane_mask() & BIT(plane_id);
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c
> > index 5dc594eafaf2..4e66261bf5af 100644
> > --- a/drivers/gpu/drm/i915/intel_tv.c
> > +++ b/drivers/gpu/drm/i915/intel_tv.c
> > @@ -906,7 +906,7 @@ static struct intel_tv *intel_attached_tv(struct drm_connector *connector)
> >   static bool
> >   intel_tv_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 tmp = I915_READ(TV_CTL);
> >   	*pipe = (tmp & TV_ENC_PIPE_SEL_MASK) >> TV_ENC_PIPE_SEL_SHIFT;
> > @@ -920,10 +920,10 @@ intel_enable_tv(struct intel_encoder *encoder,
> >   		const struct drm_connector_state *conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	/* Prevents vblank waits from timing out in intel_tv_detect_type() */
> > -	intel_wait_for_vblank(dev_priv,
> > +	intel_wait_for_vblank(i915,
> >   			      to_intel_crtc(pipe_config->base.crtc)->pipe);
> >   	I915_WRITE(TV_CTL, I915_READ(TV_CTL) | TV_ENC_ENABLE);
> > @@ -935,7 +935,7 @@ intel_disable_tv(struct intel_encoder *encoder,
> >   		 const struct drm_connector_state *old_conn_state)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	I915_WRITE(TV_CTL, I915_READ(TV_CTL) & ~TV_ENC_ENABLE);
> >   }
> > @@ -1084,7 +1084,7 @@ static void
> >   intel_tv_get_config(struct intel_encoder *encoder,
> >   		    struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct drm_display_mode *adjusted_mode =
> >   		&pipe_config->base.adjusted_mode;
> >   	struct drm_display_mode mode = {};
> > @@ -1158,15 +1158,15 @@ intel_tv_get_config(struct intel_encoder *encoder,
> >   		adjusted_mode->crtc_clock /= 2;
> >   	/* pixel counter doesn't work on i965gm TV output */
> > -	if (IS_I965GM(dev_priv))
> > +	if (IS_I965GM(i915))
> >   		adjusted_mode->private_flags |=
> >   			I915_MODE_FLAG_USE_SCANLINE_COUNTER;
> >   }
> > -static bool intel_tv_source_too_wide(struct drm_i915_private *dev_priv,
> > +static bool intel_tv_source_too_wide(struct drm_i915_private *i915,
> >   				     int hdisplay)
> >   {
> > -	return IS_GEN(dev_priv, 3) && hdisplay > 1024;
> > +	return IS_GEN(i915, 3) && hdisplay > 1024;
> >   }
> >   static bool intel_tv_vert_scaling(const struct drm_display_mode *tv_mode,
> > @@ -1184,7 +1184,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> >   			struct intel_crtc_state *pipe_config,
> >   			struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_tv_connector_state *tv_conn_state =
> >   		to_intel_tv_connector_state(conn_state);
> >   	const struct tv_mode *tv_mode = intel_tv_mode_find(conn_state);
> > @@ -1209,7 +1209,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> >   	intel_tv_mode_to_mode(adjusted_mode, tv_mode);
> >   	drm_mode_set_crtcinfo(adjusted_mode, 0);
> > -	if (intel_tv_source_too_wide(dev_priv, hdisplay) ||
> > +	if (intel_tv_source_too_wide(i915, hdisplay) ||
> >   	    !intel_tv_vert_scaling(adjusted_mode, conn_state, vdisplay)) {
> >   		int extra, top, bottom;
> > @@ -1306,7 +1306,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> >   	 * the active portion. Hence following this formula seems
> >   	 * more trouble that it's worth.
> >   	 *
> > -	 * if (IS_GEN(dev_priv, 4)) {
> > +	 * if (IS_GEN(i915, 4)) {
> >   	 *	num = cdclk * (tv_mode->oversample >> !tv_mode->progressive);
> >   	 *	den = tv_mode->clock;
> >   	 * } else {
> > @@ -1327,7 +1327,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> >   	adjusted_mode->name[0] = '\0';
> >   	/* pixel counter doesn't work on i965gm TV output */
> > -	if (IS_I965GM(dev_priv))
> > +	if (IS_I965GM(i915))
> >   		adjusted_mode->private_flags |=
> >   			I915_MODE_FLAG_USE_SCANLINE_COUNTER;
> > @@ -1335,7 +1335,7 @@ intel_tv_compute_config(struct intel_encoder *encoder,
> >   }
> >   static void
> > -set_tv_mode_timings(struct drm_i915_private *dev_priv,
> > +set_tv_mode_timings(struct drm_i915_private *i915,
> >   		    const struct tv_mode *tv_mode,
> >   		    bool burst_ena)
> >   {
> > @@ -1393,7 +1393,7 @@ set_tv_mode_timings(struct drm_i915_private *dev_priv,
> >   	I915_WRITE(TV_V_CTL_7, vctl7);
> >   }
> > -static void set_color_conversion(struct drm_i915_private *dev_priv,
> > +static void set_color_conversion(struct drm_i915_private *i915,
> >   				 const struct color_conversion *color_conversion)
> >   {
> >   	if (!color_conversion)
> > @@ -1417,7 +1417,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> >   				const struct intel_crtc_state *pipe_config,
> >   				const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	struct intel_tv *intel_tv = enc_to_tv(encoder);
> >   	const struct intel_tv_connector_state *tv_conn_state =
> > @@ -1507,18 +1507,18 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> >   		tv_mode->dda3_inc << TV_SCDDA3_INC_SHIFT;
> >   	/* Enable two fixes for the chips that need them. */
> > -	if (IS_I915GM(dev_priv))
> > +	if (IS_I915GM(i915))
> >   		tv_ctl |= TV_ENC_C0_FIX | TV_ENC_SDP_FIX;
> > -	set_tv_mode_timings(dev_priv, tv_mode, burst_ena);
> > +	set_tv_mode_timings(i915, tv_mode, burst_ena);
> >   	I915_WRITE(TV_SC_CTL_1, scctl1);
> >   	I915_WRITE(TV_SC_CTL_2, scctl2);
> >   	I915_WRITE(TV_SC_CTL_3, scctl3);
> > -	set_color_conversion(dev_priv, color_conversion);
> > +	set_color_conversion(i915, color_conversion);
> > -	if (INTEL_GEN(dev_priv) >= 4)
> > +	if (INTEL_GEN(i915) >= 4)
> >   		I915_WRITE(TV_CLR_KNOBS, 0x00404000);
> >   	else
> >   		I915_WRITE(TV_CLR_KNOBS, 0x00606000);
> > @@ -1528,7 +1528,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder,
> >   			   ((video_levels->black << TV_BLACK_LEVEL_SHIFT) |
> >   			    (video_levels->blank << TV_BLANK_LEVEL_SHIFT)));
> > -	assert_pipe_disabled(dev_priv, intel_crtc->pipe);
> > +	assert_pipe_disabled(i915, intel_crtc->pipe);
> >   	/* Filter ctl must be set before TV_WIN_SIZE */
> >   	tv_filter_ctl = TV_AUTO_SCALE;
> > @@ -1568,18 +1568,18 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> >   	struct drm_crtc *crtc = connector->state->crtc;
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	struct drm_device *dev = connector->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 tv_ctl, save_tv_ctl;
> >   	u32 tv_dac, save_tv_dac;
> >   	int type;
> >   	/* Disable TV interrupts around load detect or we'll recurse */
> >   	if (connector->polled & DRM_CONNECTOR_POLL_HPD) {
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		i915_disable_pipestat(dev_priv, 0,
> > +		spin_lock_irq(&i915->irq_lock);
> > +		i915_disable_pipestat(i915, 0,
> >   				      PIPE_HOTPLUG_INTERRUPT_STATUS |
> >   				      PIPE_HOTPLUG_TV_INTERRUPT_STATUS);
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +		spin_unlock_irq(&i915->irq_lock);
> >   	}
> >   	save_tv_dac = tv_dac = I915_READ(TV_DAC);
> > @@ -1605,7 +1605,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> >   	 * The TV sense state should be cleared to zero on cantiga platform. Otherwise
> >   	 * the TV is misdetected. This is hardware requirement.
> >   	 */
> > -	if (IS_GM45(dev_priv))
> > +	if (IS_GM45(i915))
> >   		tv_dac &= ~(TVDAC_STATE_CHG_EN | TVDAC_A_SENSE_CTL |
> >   			    TVDAC_B_SENSE_CTL | TVDAC_C_SENSE_CTL);
> > @@ -1613,7 +1613,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> >   	I915_WRITE(TV_DAC, tv_dac);
> >   	POSTING_READ(TV_DAC);
> > -	intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> > +	intel_wait_for_vblank(i915, intel_crtc->pipe);
> >   	type = -1;
> >   	tv_dac = I915_READ(TV_DAC);
> > @@ -1643,15 +1643,15 @@ intel_tv_detect_type(struct intel_tv *intel_tv,
> >   	POSTING_READ(TV_CTL);
> >   	/* For unknown reasons the hw barfs if we don't do this vblank wait. */
> > -	intel_wait_for_vblank(dev_priv, intel_crtc->pipe);
> > +	intel_wait_for_vblank(i915, intel_crtc->pipe);
> >   	/* Restore interrupt config */
> >   	if (connector->polled & DRM_CONNECTOR_POLL_HPD) {
> > -		spin_lock_irq(&dev_priv->irq_lock);
> > -		i915_enable_pipestat(dev_priv, 0,
> > +		spin_lock_irq(&i915->irq_lock);
> > +		i915_enable_pipestat(i915, 0,
> >   				     PIPE_HOTPLUG_INTERRUPT_STATUS |
> >   				     PIPE_HOTPLUG_TV_INTERRUPT_STATUS);
> > -		spin_unlock_irq(&dev_priv->irq_lock);
> > +		spin_unlock_irq(&i915->irq_lock);
> >   	}
> >   	return type;
> > @@ -1764,7 +1764,7 @@ intel_tv_set_mode_type(struct drm_display_mode *mode,
> >   static int
> >   intel_tv_get_modes(struct drm_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->dev);
> >   	const struct tv_mode *tv_mode = intel_tv_mode_find(connector->state);
> >   	int i, count = 0;
> > @@ -1778,7 +1778,7 @@ intel_tv_get_modes(struct drm_connector *connector)
> >   			continue;
> >   		/* no vertical scaling with wide sources on gen3 */
> > -		if (IS_GEN(dev_priv, 3) && input->w > 1024 &&
> > +		if (IS_GEN(i915, 3) && input->w > 1024 &&
> >   		    input->h > intel_tv_mode_vdisplay(tv_mode))
> >   			continue;
> > @@ -1857,9 +1857,9 @@ static const struct drm_encoder_funcs intel_tv_enc_funcs = {
> >   };
> >   void
> > -intel_tv_init(struct drm_i915_private *dev_priv)
> > +intel_tv_init(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct drm_connector *connector;
> >   	struct intel_tv *intel_tv;
> >   	struct intel_encoder *intel_encoder;
> > @@ -1872,7 +1872,7 @@ intel_tv_init(struct drm_i915_private *dev_priv)
> >   	if ((I915_READ(TV_CTL) & TV_FUSE_STATE_MASK) == TV_FUSE_STATE_DISABLED)
> >   		return;
> > -	if (!intel_bios_is_tv_present(dev_priv)) {
> > +	if (!intel_bios_is_tv_present(i915)) {
> >   		DRM_DEBUG_KMS("Integrated TV is not present.\n");
> >   		return;
> >   	}
> > @@ -1966,7 +1966,7 @@ intel_tv_init(struct drm_i915_private *dev_priv)
> >   	/* Create TV properties then attach current values */
> >   	for (i = 0; i < ARRAY_SIZE(tv_modes); i++) {
> >   		/* 1080p50/1080p60 not supported on gen3 */
> > -		if (IS_GEN(dev_priv, 3) &&
> > +		if (IS_GEN(i915, 3) &&
> >   		    tv_modes[i].oversample == 1)
> >   			break;
> > diff --git a/drivers/gpu/drm/i915/intel_tv.h b/drivers/gpu/drm/i915/intel_tv.h
> > index 44518575ec5c..fc0b89563b4f 100644
> > --- a/drivers/gpu/drm/i915/intel_tv.h
> > +++ b/drivers/gpu/drm/i915/intel_tv.h
> > @@ -8,6 +8,6 @@
> >   struct drm_i915_private;
> > -void intel_tv_init(struct drm_i915_private *dev_priv);
> > +void intel_tv_init(struct drm_i915_private *i915);
> >   #endif /* __INTEL_TV_H__ */
> > diff --git a/drivers/gpu/drm/i915/intel_uc.c b/drivers/gpu/drm/i915/intel_uc.c
> > index a8e7f0ba7c3b..11e2dcf6b917 100644
> > --- a/drivers/gpu/drm/i915/intel_uc.c
> > +++ b/drivers/gpu/drm/i915/intel_uc.c
> > @@ -33,12 +33,12 @@ static void guc_free_load_err_log(struct intel_guc *guc);
> >   /* Reset GuC providing us with fresh state for both GuC and HuC.
> >    */
> > -static int __intel_uc_reset_hw(struct drm_i915_private *dev_priv)
> > +static int __intel_uc_reset_hw(struct drm_i915_private *i915)
> >   {
> >   	int ret;
> >   	u32 guc_status;
> > -	ret = intel_reset_guc(dev_priv);
> > +	ret = intel_reset_guc(i915);
> >   	if (ret) {
> >   		DRM_ERROR("Failed to reset GuC, ret = %d\n", ret);
> >   		return ret;
> > diff --git a/drivers/gpu/drm/i915/intel_uc.h b/drivers/gpu/drm/i915/intel_uc.h
> > index 3ea06c87dfcd..4ddd50d990b3 100644
> > --- a/drivers/gpu/drm/i915/intel_uc.h
> > +++ b/drivers/gpu/drm/i915/intel_uc.h
> > @@ -28,20 +28,20 @@
> >   #include "intel_huc.h"
> >   #include "i915_params.h"
> > -void intel_uc_init_early(struct drm_i915_private *dev_priv);
> > -void intel_uc_cleanup_early(struct drm_i915_private *dev_priv);
> > -void intel_uc_init_mmio(struct drm_i915_private *dev_priv);
> > -int intel_uc_init_misc(struct drm_i915_private *dev_priv);
> > -void intel_uc_fini_misc(struct drm_i915_private *dev_priv);
> > -void intel_uc_sanitize(struct drm_i915_private *dev_priv);
> > -int intel_uc_init_hw(struct drm_i915_private *dev_priv);
> > -void intel_uc_fini_hw(struct drm_i915_private *dev_priv);
> > -int intel_uc_init(struct drm_i915_private *dev_priv);
> > -void intel_uc_fini(struct drm_i915_private *dev_priv);
> > +void intel_uc_init_early(struct drm_i915_private *i915);
> > +void intel_uc_cleanup_early(struct drm_i915_private *i915);
> > +void intel_uc_init_mmio(struct drm_i915_private *i915);
> > +int intel_uc_init_misc(struct drm_i915_private *i915);
> > +void intel_uc_fini_misc(struct drm_i915_private *i915);
> > +void intel_uc_sanitize(struct drm_i915_private *i915);
> > +int intel_uc_init_hw(struct drm_i915_private *i915);
> > +void intel_uc_fini_hw(struct drm_i915_private *i915);
> > +int intel_uc_init(struct drm_i915_private *i915);
> > +void intel_uc_fini(struct drm_i915_private *i915);
> >   void intel_uc_reset_prepare(struct drm_i915_private *i915);
> >   void intel_uc_suspend(struct drm_i915_private *i915);
> >   void intel_uc_runtime_suspend(struct drm_i915_private *i915);
> > -int intel_uc_resume(struct drm_i915_private *dev_priv);
> > +int intel_uc_resume(struct drm_i915_private *i915);
> >   static inline bool intel_uc_is_using_guc(struct drm_i915_private *i915)
> >   {
> > diff --git a/drivers/gpu/drm/i915/intel_uc_fw.c b/drivers/gpu/drm/i915/intel_uc_fw.c
> > index f342ddd47df8..d156911d1560 100644
> > --- a/drivers/gpu/drm/i915/intel_uc_fw.c
> > +++ b/drivers/gpu/drm/i915/intel_uc_fw.c
> > @@ -32,15 +32,15 @@
> >   /**
> >    * intel_uc_fw_fetch - fetch uC firmware
> >    *
> > - * @dev_priv: device private
> > + * @i915: device private
> >    * @uc_fw: uC firmware
> >    *
> >    * Fetch uC firmware into GEM obj.
> >    */
> > -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> > +void intel_uc_fw_fetch(struct drm_i915_private *i915,
> >   		       struct intel_uc_fw *uc_fw)
> >   {
> > -	struct pci_dev *pdev = dev_priv->drm.pdev;
> > +	struct pci_dev *pdev = i915->drm.pdev;
> >   	struct drm_i915_gem_object *obj;
> >   	const struct firmware *fw = NULL;
> >   	struct uc_css_header *css;
> > @@ -48,10 +48,10 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> >   	int err;
> >   	if (!uc_fw->path) {
> > -		dev_info(dev_priv->drm.dev,
> > +		dev_info(i915->drm.dev,
> >   			 "%s: No firmware was defined for %s!\n",
> >   			 intel_uc_fw_type_repr(uc_fw->type),
> > -			 intel_platform_name(INTEL_INFO(dev_priv)->platform));
> > +			 intel_platform_name(INTEL_INFO(i915)->platform));
> >   		return;
> >   	}
> > @@ -159,7 +159,7 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> >   		goto fail;
> >   	}
> > -	obj = i915_gem_object_create_shmem_from_data(dev_priv,
> > +	obj = i915_gem_object_create_shmem_from_data(i915,
> >   						     fw->data, fw->size);
> >   	if (IS_ERR(obj)) {
> >   		err = PTR_ERR(obj);
> > diff --git a/drivers/gpu/drm/i915/intel_uc_fw.h b/drivers/gpu/drm/i915/intel_uc_fw.h
> > index ff98f8661d72..881c3aef074c 100644
> > --- a/drivers/gpu/drm/i915/intel_uc_fw.h
> > +++ b/drivers/gpu/drm/i915/intel_uc_fw.h
> > @@ -142,7 +142,7 @@ static inline u32 intel_uc_fw_get_upload_size(struct intel_uc_fw *uc_fw)
> >   	return uc_fw->header_size + uc_fw->ucode_size;
> >   }
> > -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv,
> > +void intel_uc_fw_fetch(struct drm_i915_private *i915,
> >   		       struct intel_uc_fw *uc_fw);
> >   void intel_uc_fw_cleanup_fetch(struct intel_uc_fw *uc_fw);
> >   int intel_uc_fw_upload(struct intel_uc_fw *uc_fw,
> > diff --git a/drivers/gpu/drm/i915/intel_uncore.c b/drivers/gpu/drm/i915/intel_uncore.c
> > index 85171a8b866a..87dec005165d 100644
> > --- a/drivers/gpu/drm/i915/intel_uncore.c
> > +++ b/drivers/gpu/drm/i915/intel_uncore.c
> > @@ -537,10 +537,10 @@ void intel_uncore_runtime_resume(struct intel_uncore *uncore)
> >   	iosf_mbi_register_pmic_bus_access_notifier(&uncore->pmic_bus_access_nb);
> >   }
> > -void intel_uncore_sanitize(struct drm_i915_private *dev_priv)
> > +void intel_uncore_sanitize(struct drm_i915_private *i915)
> >   {
> >   	/* BIOS often leaves RC6 enabled, but disable it for hw init */
> > -	intel_sanitize_gt_powersave(dev_priv);
> > +	intel_sanitize_gt_powersave(i915);
> >   }
> >   static void __intel_uncore_forcewake_get(struct intel_uncore *uncore,
> > @@ -647,7 +647,7 @@ void intel_uncore_forcewake_user_put(struct intel_uncore *uncore)
> >    * @fw_domains: forcewake domains to get reference on
> >    *
> >    * See intel_uncore_forcewake_get(). This variant places the onus
> > - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock.
> > + * on the caller to explicitly handle the i915->uncore.lock spinlock.
> >    */
> >   void intel_uncore_forcewake_get__locked(struct intel_uncore *uncore,
> >   					enum forcewake_domains fw_domains)
> > @@ -708,7 +708,7 @@ void intel_uncore_forcewake_put(struct intel_uncore *uncore,
> >    * @fw_domains: forcewake domains to get reference on
> >    *
> >    * See intel_uncore_forcewake_put(). This variant places the onus
> > - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock.
> > + * on the caller to explicitly handle the i915->uncore.lock spinlock.
> >    */
> >   void intel_uncore_forcewake_put__locked(struct intel_uncore *uncore,
> >   					enum forcewake_domains fw_domains)
> > @@ -821,10 +821,10 @@ find_fw_domain(struct intel_uncore *uncore, u32 offset)
> >   #define GEN_FW_RANGE(s, e, d) \
> >   	{ .start = (s), .end = (e), .domains = (d) }
> > -#define HAS_FWTABLE(dev_priv) \
> > -	(INTEL_GEN(dev_priv) >= 9 || \
> > -	 IS_CHERRYVIEW(dev_priv) || \
> > -	 IS_VALLEYVIEW(dev_priv))
> > +#define HAS_FWTABLE(i915) \
> > +	(INTEL_GEN(i915) >= 9 || \
> > +	 IS_CHERRYVIEW(i915) || \
> > +	 IS_VALLEYVIEW(i915))
> >   /* *Must* be sorted by offset ranges! See intel_fw_table_check(). */
> >   static const struct intel_forcewake_range __vlv_fw_ranges[] = {
> > @@ -1461,7 +1461,7 @@ static void intel_uncore_fw_domains_init(struct intel_uncore *uncore)
> >   static int i915_pmic_bus_access_notifier(struct notifier_block *nb,
> >   					 unsigned long action, void *data)
> >   {
> > -	struct drm_i915_private *dev_priv = container_of(nb,
> > +	struct drm_i915_private *i915 = container_of(nb,
> >   			struct drm_i915_private, uncore.pmic_bus_access_nb);
> >   	switch (action) {
> > @@ -1479,12 +1479,12 @@ static int i915_pmic_bus_access_notifier(struct notifier_block *nb,
> >   		 * wake reference -> disable wakeref asserts for the time of
> >   		 * the access.
> >   		 */
> > -		disable_rpm_wakeref_asserts(dev_priv);
> > -		intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL);
> > -		enable_rpm_wakeref_asserts(dev_priv);
> > +		disable_rpm_wakeref_asserts(i915);
> > +		intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL);
> > +		enable_rpm_wakeref_asserts(i915);
> >   		break;
> >   	case MBI_PMIC_BUS_ACCESS_END:
> > -		intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL);
> > +		intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL);
> >   		break;
> >   	}
> > diff --git a/drivers/gpu/drm/i915/intel_uncore.h b/drivers/gpu/drm/i915/intel_uncore.h
> > index d6af3de70121..d30bda2c3444 100644
> > --- a/drivers/gpu/drm/i915/intel_uncore.h
> > +++ b/drivers/gpu/drm/i915/intel_uncore.h
> > @@ -177,7 +177,7 @@ intel_uncore_has_fifo(const struct intel_uncore *uncore)
> >   	return uncore->flags & UNCORE_HAS_FIFO;
> >   }
> > -void intel_uncore_sanitize(struct drm_i915_private *dev_priv);
> > +void intel_uncore_sanitize(struct drm_i915_private *i915);
> >   void intel_uncore_init_early(struct intel_uncore *uncore);
> >   int intel_uncore_init_mmio(struct intel_uncore *uncore);
> >   void intel_uncore_prune_mmio_domains(struct intel_uncore *uncore);
> > @@ -361,7 +361,7 @@ intel_uncore_read64_2x32(struct intel_uncore *uncore,
> >    *
> >    * Certain architectures will die if the same cacheline is concurrently accessed
> >    * by different clients (e.g. on Ivybridge). Access to registers should
> > - * therefore generally be serialised, by either the dev_priv->uncore.lock or
> > + * therefore generally be serialised, by either the i915->uncore.lock or
> >    * a more localised lock guarding all access to that bank of registers.
> >    */
> >   #define intel_uncore_read_fw(...) __raw_uncore_read32(__VA_ARGS__)
> > diff --git a/drivers/gpu/drm/i915/intel_vbt_defs.h b/drivers/gpu/drm/i915/intel_vbt_defs.h
> > index 89ef14cafb6b..e2e9883d6b28 100644
> > --- a/drivers/gpu/drm/i915/intel_vbt_defs.h
> > +++ b/drivers/gpu/drm/i915/intel_vbt_defs.h
> > @@ -337,7 +337,7 @@ enum vbt_gmbus_ddi {
> >    * basically any of the fields to ensure the correct interpretation for the BDB
> >    * version in question.
> >    *
> > - * When we copy the child device configs to dev_priv->vbt.child_dev, we reserve
> > + * When we copy the child device configs to i915->vbt.child_dev, we reserve
> >    * space for the full structure below, and initialize the tail not actually
> >    * present in VBT to zeros. Accessing those fields is fine, as long as the
> >    * default zero is taken into account, again according to the BDB version.
> > diff --git a/drivers/gpu/drm/i915/intel_vdsc.c b/drivers/gpu/drm/i915/intel_vdsc.c
> > index ffec807b8960..164fd4b56112 100644
> > --- a/drivers/gpu/drm/i915/intel_vdsc.c
> > +++ b/drivers/gpu/drm/i915/intel_vdsc.c
> > @@ -478,7 +478,7 @@ static void intel_configure_pps_for_dsc_encoder(struct intel_encoder *encoder,
> >   						const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	const struct drm_dsc_config *vdsc_cfg = &crtc_state->dp_dsc_cfg;
> >   	enum pipe pipe = crtc->pipe;
> >   	enum transcoder cpu_transcoder = crtc_state->cpu_transcoder;
> > @@ -897,7 +897,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> >   		      const struct intel_crtc_state *crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	i915_reg_t dss_ctl1_reg, dss_ctl2_reg;
> >   	u32 dss_ctl1_val = 0;
> > @@ -907,7 +907,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> >   		return;
> >   	/* Enable Power wells for VDSC/joining */
> > -	intel_display_power_get(dev_priv,
> > +	intel_display_power_get(i915,
> >   				intel_dsc_power_domain(crtc_state));
> >   	intel_configure_pps_for_dsc_encoder(encoder, crtc_state);
> > @@ -933,7 +933,7 @@ void intel_dsc_enable(struct intel_encoder *encoder,
> >   void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state)
> >   {
> >   	struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc);
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
> >   	enum pipe pipe = crtc->pipe;
> >   	i915_reg_t dss_ctl1_reg, dss_ctl2_reg;
> >   	u32 dss_ctl1_val = 0, dss_ctl2_val = 0;
> > @@ -961,6 +961,6 @@ void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state)
> >   	I915_WRITE(dss_ctl2_reg, dss_ctl2_val);
> >   	/* Disable Power wells for VDSC/joining */
> > -	intel_display_power_put_unchecked(dev_priv,
> > +	intel_display_power_put_unchecked(i915,
> >   					  intel_dsc_power_domain(old_crtc_state));
> >   }
> > diff --git a/drivers/gpu/drm/i915/intel_wopcm.c b/drivers/gpu/drm/i915/intel_wopcm.c
> > index 7b4ba84b9fb8..646540b97267 100644
> > --- a/drivers/gpu/drm/i915/intel_wopcm.c
> > +++ b/drivers/gpu/drm/i915/intel_wopcm.c
> > @@ -225,7 +225,7 @@ int intel_wopcm_init(struct intel_wopcm *wopcm)
> >   	return 0;
> >   }
> > -static inline int write_and_verify(struct drm_i915_private *dev_priv,
> > +static inline int write_and_verify(struct drm_i915_private *i915,
> >   				   i915_reg_t reg, u32 val, u32 mask,
> >   				   u32 locked_bit)
> >   {
> > @@ -252,27 +252,27 @@ static inline int write_and_verify(struct drm_i915_private *dev_priv,
> >    */
> >   int intel_wopcm_init_hw(struct intel_wopcm *wopcm)
> >   {
> > -	struct drm_i915_private *dev_priv = wopcm_to_i915(wopcm);
> > +	struct drm_i915_private *i915 = wopcm_to_i915(wopcm);
> >   	u32 huc_agent;
> >   	u32 mask;
> >   	int err;
> > -	if (!USES_GUC(dev_priv))
> > +	if (!USES_GUC(i915))
> >   		return 0;
> > -	GEM_BUG_ON(!HAS_GUC(dev_priv));
> > +	GEM_BUG_ON(!HAS_GUC(i915));
> >   	GEM_BUG_ON(!wopcm->guc.size);
> >   	GEM_BUG_ON(!wopcm->guc.base);
> > -	err = write_and_verify(dev_priv, GUC_WOPCM_SIZE, wopcm->guc.size,
> > +	err = write_and_verify(i915, GUC_WOPCM_SIZE, wopcm->guc.size,
> >   			       GUC_WOPCM_SIZE_MASK | GUC_WOPCM_SIZE_LOCKED,
> >   			       GUC_WOPCM_SIZE_LOCKED);
> >   	if (err)
> >   		goto err_out;
> > -	huc_agent = USES_HUC(dev_priv) ? HUC_LOADING_AGENT_GUC : 0;
> > +	huc_agent = USES_HUC(i915) ? HUC_LOADING_AGENT_GUC : 0;
> >   	mask = GUC_WOPCM_OFFSET_MASK | GUC_WOPCM_OFFSET_VALID | huc_agent;
> > -	err = write_and_verify(dev_priv, DMA_GUC_WOPCM_OFFSET,
> > +	err = write_and_verify(i915, DMA_GUC_WOPCM_OFFSET,
> >   			       wopcm->guc.base | huc_agent, mask,
> >   			       GUC_WOPCM_OFFSET_VALID);
> >   	if (err)
> > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> > index 2093d08a7569..2514842236f6 100644
> > --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> > @@ -147,17 +147,17 @@ fake_dma_object(struct drm_i915_private *i915, u64 size)
> >   static int igt_ppgtt_alloc(void *arg)
> >   {
> > -	struct drm_i915_private *dev_priv = arg;
> > +	struct drm_i915_private *i915 = arg;
> >   	struct i915_ppgtt *ppgtt;
> >   	u64 size, last, limit;
> >   	int err = 0;
> >   	/* Allocate a ppggt and try to fill the entire range */
> > -	if (!HAS_PPGTT(dev_priv))
> > +	if (!HAS_PPGTT(i915))
> >   		return 0;
> > -	ppgtt = __ppgtt_create(dev_priv);
> > +	ppgtt = __ppgtt_create(i915);
> >   	if (IS_ERR(ppgtt))
> >   		return PTR_ERR(ppgtt);
> > @@ -208,9 +208,9 @@ static int igt_ppgtt_alloc(void *arg)
> >   	}
> >   err_ppgtt_cleanup:
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > +	mutex_lock(&i915->drm.struct_mutex);
> >   	i915_vm_put(&ppgtt->vm);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return err;
> >   }
> > @@ -992,7 +992,7 @@ static int shrink_boom(struct drm_i915_private *i915,
> >   	return err;
> >   }
> > -static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> > +static int exercise_ppgtt(struct drm_i915_private *i915,
> >   			  int (*func)(struct drm_i915_private *i915,
> >   				      struct i915_address_space *vm,
> >   				      u64 hole_start, u64 hole_end,
> > @@ -1003,15 +1003,15 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> >   	IGT_TIMEOUT(end_time);
> >   	int err;
> > -	if (!HAS_FULL_PPGTT(dev_priv))
> > +	if (!HAS_FULL_PPGTT(i915))
> >   		return 0;
> > -	file = mock_file(dev_priv);
> > +	file = mock_file(i915);
> >   	if (IS_ERR(file))
> >   		return PTR_ERR(file);
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	ppgtt = i915_ppgtt_create(dev_priv);
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	ppgtt = i915_ppgtt_create(i915);
> >   	if (IS_ERR(ppgtt)) {
> >   		err = PTR_ERR(ppgtt);
> >   		goto out_unlock;
> > @@ -1019,13 +1019,13 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv,
> >   	GEM_BUG_ON(offset_in_page(ppgtt->vm.total));
> >   	GEM_BUG_ON(ppgtt->vm.closed);
> > -	err = func(dev_priv, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
> > +	err = func(i915, &ppgtt->vm, 0, ppgtt->vm.total, end_time);
> >   	i915_vm_put(&ppgtt->vm);
> >   out_unlock:
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> > -	mock_file_free(dev_priv, file);
> > +	mock_file_free(i915, file);
> >   	return err;
> >   }
> > diff --git a/drivers/gpu/drm/i915/selftests/intel_guc.c b/drivers/gpu/drm/i915/selftests/intel_guc.c
> > index 7fd0321e0947..ab3e71c957df 100644
> > --- a/drivers/gpu/drm/i915/selftests/intel_guc.c
> > +++ b/drivers/gpu/drm/i915/selftests/intel_guc.c
> > @@ -107,12 +107,12 @@ static int validate_client(struct intel_guc_client *client,
> >   			   int client_priority,
> >   			   bool is_preempt_client)
> >   {
> > -	struct drm_i915_private *dev_priv = guc_to_i915(client->guc);
> > +	struct drm_i915_private *i915 = guc_to_i915(client->guc);
> >   	struct i915_gem_context *ctx_owner = is_preempt_client ?
> > -			dev_priv->preempt_context : dev_priv->kernel_context;
> > +			i915->preempt_context : i915->kernel_context;
> >   	if (client->owner != ctx_owner ||
> > -	    client->engines != INTEL_INFO(dev_priv)->engine_mask ||
> > +	    client->engines != INTEL_INFO(i915)->engine_mask ||
> >   	    client->priority != client_priority ||
> >   	    client->doorbell_id == GUC_DOORBELL_INVALID)
> >   		return -EINVAL;
> > @@ -137,16 +137,16 @@ static bool client_doorbell_in_sync(struct intel_guc_client *client)
> >    */
> >   static int igt_guc_clients(void *args)
> >   {
> > -	struct drm_i915_private *dev_priv = args;
> > +	struct drm_i915_private *i915 = args;
> >   	intel_wakeref_t wakeref;
> >   	struct intel_guc *guc;
> >   	int err = 0;
> > -	GEM_BUG_ON(!HAS_GUC(dev_priv));
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	GEM_BUG_ON(!HAS_GUC(i915));
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	guc = &dev_priv->guc;
> > +	guc = &i915->guc;
> >   	if (!guc) {
> >   		pr_err("No guc object!\n");
> >   		err = -EINVAL;
> > @@ -227,8 +227,8 @@ static int igt_guc_clients(void *args)
> >   	guc_clients_create(guc);
> >   	guc_clients_enable(guc);
> >   unlock:
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	intel_runtime_pm_put(i915, wakeref);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return err;
> >   }
> > @@ -239,17 +239,17 @@ static int igt_guc_clients(void *args)
> >    */
> >   static int igt_guc_doorbells(void *arg)
> >   {
> > -	struct drm_i915_private *dev_priv = arg;
> > +	struct drm_i915_private *i915 = arg;
> >   	intel_wakeref_t wakeref;
> >   	struct intel_guc *guc;
> >   	int i, err = 0;
> >   	u16 db_id;
> > -	GEM_BUG_ON(!HAS_GUC(dev_priv));
> > -	mutex_lock(&dev_priv->drm.struct_mutex);
> > -	wakeref = intel_runtime_pm_get(dev_priv);
> > +	GEM_BUG_ON(!HAS_GUC(i915));
> > +	mutex_lock(&i915->drm.struct_mutex);
> > +	wakeref = intel_runtime_pm_get(i915);
> > -	guc = &dev_priv->guc;
> > +	guc = &i915->guc;
> >   	if (!guc) {
> >   		pr_err("No guc object!\n");
> >   		err = -EINVAL;
> > @@ -261,10 +261,10 @@ static int igt_guc_doorbells(void *arg)
> >   		goto unlock;
> >   	for (i = 0; i < ATTEMPTS; i++) {
> > -		clients[i] = guc_client_alloc(dev_priv,
> > -					      INTEL_INFO(dev_priv)->engine_mask,
> > +		clients[i] = guc_client_alloc(i915,
> > +					      INTEL_INFO(i915)->engine_mask,
> >   					      i % GUC_CLIENT_PRIORITY_NUM,
> > -					      dev_priv->kernel_context);
> > +					      i915->kernel_context);
> >   		if (!clients[i]) {
> >   			pr_err("[%d] No guc client\n", i);
> > @@ -340,20 +340,20 @@ static int igt_guc_doorbells(void *arg)
> >   			guc_client_free(clients[i]);
> >   		}
> >   unlock:
> > -	intel_runtime_pm_put(dev_priv, wakeref);
> > -	mutex_unlock(&dev_priv->drm.struct_mutex);
> > +	intel_runtime_pm_put(i915, wakeref);
> > +	mutex_unlock(&i915->drm.struct_mutex);
> >   	return err;
> >   }
> > -int intel_guc_live_selftest(struct drm_i915_private *dev_priv)
> > +int intel_guc_live_selftest(struct drm_i915_private *i915)
> >   {
> >   	static const struct i915_subtest tests[] = {
> >   		SUBTEST(igt_guc_clients),
> >   		SUBTEST(igt_guc_doorbells),
> >   	};
> > -	if (!USES_GUC_SUBMISSION(dev_priv))
> > +	if (!USES_GUC_SUBMISSION(i915))
> >   		return 0;
> > -	return i915_subtests(tests, dev_priv);
> > +	return i915_subtests(tests, i915);
> >   }
> > diff --git a/drivers/gpu/drm/i915/selftests/intel_uncore.c b/drivers/gpu/drm/i915/selftests/intel_uncore.c
> > index e0d7ebecb215..b5304cfeda21 100644
> > --- a/drivers/gpu/drm/i915/selftests/intel_uncore.c
> > +++ b/drivers/gpu/drm/i915/selftests/intel_uncore.c
> > @@ -254,15 +254,15 @@ static int live_forcewake_ops(void *arg)
> >   static int live_forcewake_domains(void *arg)
> >   {
> >   #define FW_RANGE 0x40000
> > -	struct drm_i915_private *dev_priv = arg;
> > -	struct intel_uncore *uncore = &dev_priv->uncore;
> > +	struct drm_i915_private *i915 = arg;
> > +	struct intel_uncore *uncore = &i915->uncore;
> >   	unsigned long *valid;
> >   	u32 offset;
> >   	int err;
> > -	if (!HAS_FPGA_DBG_UNCLAIMED(dev_priv) &&
> > -	    !IS_VALLEYVIEW(dev_priv) &&
> > -	    !IS_CHERRYVIEW(dev_priv))
> > +	if (!HAS_FPGA_DBG_UNCLAIMED(i915) &&
> > +	    !IS_VALLEYVIEW(i915) &&
> > +	    !IS_CHERRYVIEW(i915))
> >   		return 0;
> >   	/*
> > diff --git a/drivers/gpu/drm/i915/vlv_dsi.c b/drivers/gpu/drm/i915/vlv_dsi.c
> > index e272d826210a..0178bdf0c5a9 100644
> > --- a/drivers/gpu/drm/i915/vlv_dsi.c
> > +++ b/drivers/gpu/drm/i915/vlv_dsi.c
> > @@ -78,19 +78,19 @@ void vlv_dsi_wait_for_fifo_empty(struct intel_dsi *intel_dsi, enum port port)
> >   {
> >   	struct drm_encoder *encoder = &intel_dsi->base.base;
> >   	struct drm_device *dev = encoder->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 mask;
> >   	mask = LP_CTRL_FIFO_EMPTY | HS_CTRL_FIFO_EMPTY |
> >   		LP_DATA_FIFO_EMPTY | HS_DATA_FIFO_EMPTY;
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    MIPI_GEN_FIFO_STAT(port), mask, mask,
> >   				    100))
> >   		DRM_ERROR("DPI FIFOs are not empty\n");
> >   }
> > -static void write_data(struct drm_i915_private *dev_priv,
> > +static void write_data(struct drm_i915_private *i915,
> >   		       i915_reg_t reg,
> >   		       const u8 *data, u32 len)
> >   {
> > @@ -106,7 +106,7 @@ static void write_data(struct drm_i915_private *dev_priv,
> >   	}
> >   }
> > -static void read_data(struct drm_i915_private *dev_priv,
> > +static void read_data(struct drm_i915_private *i915,
> >   		      i915_reg_t reg,
> >   		      u8 *data, u32 len)
> >   {
> > @@ -125,7 +125,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> >   {
> >   	struct intel_dsi_host *intel_dsi_host = to_intel_dsi_host(host);
> >   	struct drm_device *dev = intel_dsi_host->intel_dsi->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	enum port port = intel_dsi_host->port;
> >   	struct mipi_dsi_packet packet;
> >   	ssize_t ret;
> > @@ -154,13 +154,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> >   	/* note: this is never true for reads */
> >   	if (packet.payload_length) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_GEN_FIFO_STAT(port),
> >   					    data_mask, 0,
> >   					    50))
> >   			DRM_ERROR("Timeout waiting for HS/LP DATA FIFO !full\n");
> > -		write_data(dev_priv, data_reg, packet.payload,
> > +		write_data(i915, data_reg, packet.payload,
> >   			   packet.payload_length);
> >   	}
> > @@ -168,7 +168,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> >   		I915_WRITE(MIPI_INTR_STAT(port), GEN_READ_DATA_AVAIL);
> >   	}
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    MIPI_GEN_FIFO_STAT(port),
> >   				    ctrl_mask, 0,
> >   				    50)) {
> > @@ -180,13 +180,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host,
> >   	/* ->rx_len is set only for reads */
> >   	if (msg->rx_len) {
> >   		data_mask = GEN_READ_DATA_AVAIL;
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_INTR_STAT(port),
> >   					    data_mask, data_mask,
> >   					    50))
> >   			DRM_ERROR("Timeout waiting for read data.\n");
> > -		read_data(dev_priv, data_reg, msg->rx_buf, msg->rx_len);
> > +		read_data(i915, data_reg, msg->rx_buf, msg->rx_len);
> >   	}
> >   	/* XXX: fix for reads and writes */
> > @@ -221,7 +221,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> >   {
> >   	struct drm_encoder *encoder = &intel_dsi->base.base;
> >   	struct drm_device *dev = encoder->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 mask;
> >   	/* XXX: pipe, hs */
> > @@ -240,7 +240,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> >   	I915_WRITE(MIPI_DPI_CONTROL(port), cmd);
> >   	mask = SPL_PKT_SENT_INTERRUPT;
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    MIPI_INTR_STAT(port), mask, mask,
> >   				    100))
> >   		DRM_ERROR("Video mode command 0x%08x send failed.\n", cmd);
> > @@ -248,25 +248,25 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs,
> >   	return 0;
> >   }
> > -static void band_gap_reset(struct drm_i915_private *dev_priv)
> > +static void band_gap_reset(struct drm_i915_private *i915)
> >   {
> > -	vlv_flisdsi_get(dev_priv);
> > +	vlv_flisdsi_get(i915);
> > -	vlv_flisdsi_write(dev_priv, 0x08, 0x0001);
> > -	vlv_flisdsi_write(dev_priv, 0x0F, 0x0005);
> > -	vlv_flisdsi_write(dev_priv, 0x0F, 0x0025);
> > +	vlv_flisdsi_write(i915, 0x08, 0x0001);
> > +	vlv_flisdsi_write(i915, 0x0F, 0x0005);
> > +	vlv_flisdsi_write(i915, 0x0F, 0x0025);
> >   	udelay(150);
> > -	vlv_flisdsi_write(dev_priv, 0x0F, 0x0000);
> > -	vlv_flisdsi_write(dev_priv, 0x08, 0x0000);
> > +	vlv_flisdsi_write(i915, 0x0F, 0x0000);
> > +	vlv_flisdsi_write(i915, 0x08, 0x0000);
> > -	vlv_flisdsi_put(dev_priv);
> > +	vlv_flisdsi_put(i915);
> >   }
> >   static int intel_dsi_compute_config(struct intel_encoder *encoder,
> >   				    struct intel_crtc_state *pipe_config,
> >   				    struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = container_of(encoder, struct intel_dsi,
> >   						   base);
> >   	struct intel_connector *intel_connector = intel_dsi->attached_connector;
> > @@ -281,7 +281,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
> >   	if (fixed_mode) {
> >   		intel_fixed_panel_mode(fixed_mode, adjusted_mode);
> > -		if (HAS_GMCH(dev_priv))
> > +		if (HAS_GMCH(i915))
> >   			intel_gmch_panel_fitting(crtc, pipe_config,
> >   						 conn_state->scaling_mode);
> >   		else
> > @@ -300,7 +300,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
> >   	else
> >   		pipe_config->pipe_bpp = 18;
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		/* Enable Frame time stamp based scanline reporting */
> >   		adjusted_mode->private_flags |=
> >   			I915_MODE_FLAG_GET_SCANLINE_FROM_TIMESTAMP;
> > @@ -327,7 +327,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder,
> >   static bool glk_dsi_enable_io(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -359,7 +359,7 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder)
> >   	/* Wait for Pwr ACK */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_MIPIIO_PORT_POWERED,
> >   					    GLK_MIPIIO_PORT_POWERED,
> > @@ -378,14 +378,14 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder)
> >   static void glk_dsi_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> >   	/* Wait for MIPI PHY status bit to set */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_PHY_STATUS_PORT_READY,
> >   					    GLK_PHY_STATUS_PORT_READY,
> > @@ -413,7 +413,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
> >   			I915_WRITE(MIPI_DEVICE_READY(port), val);
> >   			/* Wait for ULPS active */
> > -			if (intel_wait_for_register(&dev_priv->uncore,
> > +			if (intel_wait_for_register(&i915->uncore,
> >   						    MIPI_CTRL(port),
> >   						    GLK_ULPS_NOT_ACTIVE,
> >   						    0,
> > @@ -440,7 +440,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
> >   	/* Wait for Stop state */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_DATA_LANE_STOP_STATE,
> >   					    GLK_DATA_LANE_STOP_STATE,
> > @@ -450,7 +450,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
> >   	/* Wait for AFE LATCH */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    BXT_MIPI_PORT_CTRL(port),
> >   					    AFE_LATCHOUT,
> >   					    AFE_LATCHOUT,
> > @@ -461,7 +461,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder)
> >   static void bxt_dsi_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> > @@ -488,21 +488,21 @@ static void bxt_dsi_device_ready(struct intel_encoder *encoder)
> >   static void vlv_dsi_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> >   	DRM_DEBUG_KMS("\n");
> > -	vlv_flisdsi_get(dev_priv);
> > +	vlv_flisdsi_get(i915);
> >   	/* program rcomp for compliance, reduce from 50 ohms to 45 ohms
> >   	 * needed everytime after power gate */
> > -	vlv_flisdsi_write(dev_priv, 0x04, 0x0004);
> > -	vlv_flisdsi_put(dev_priv);
> > +	vlv_flisdsi_write(i915, 0x04, 0x0004);
> > +	vlv_flisdsi_put(i915);
> >   	/* bandgap reset is needed after everytime we do power gate */
> > -	band_gap_reset(dev_priv);
> > +	band_gap_reset(i915);
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > @@ -527,11 +527,11 @@ static void vlv_dsi_device_ready(struct intel_encoder *encoder)
> >   static void intel_dsi_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > -	if (IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEMINILAKE(i915))
> >   		glk_dsi_device_ready(encoder);
> > -	else if (IS_GEN9_LP(dev_priv))
> > +	else if (IS_GEN9_LP(i915))
> >   		bxt_dsi_device_ready(encoder);
> >   	else
> >   		vlv_dsi_device_ready(encoder);
> > @@ -539,7 +539,7 @@ static void intel_dsi_device_ready(struct intel_encoder *encoder)
> >   static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> > @@ -554,7 +554,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
> >   	/* Wait for MIPI PHY status bit to unset */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_PHY_STATUS_PORT_READY, 0, 20))
> >   			DRM_ERROR("PHY is not turning OFF\n");
> > @@ -562,7 +562,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
> >   	/* Wait for Pwr ACK bit to unset */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_MIPIIO_PORT_POWERED, 0, 20))
> >   			DRM_ERROR("MIPI IO Port is not powergated\n");
> > @@ -571,7 +571,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder)
> >   static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 tmp;
> > @@ -583,7 +583,7 @@ static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder)
> >   	/* Wait for MIPI PHY status bit to unset */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (intel_wait_for_register(&dev_priv->uncore,
> > +		if (intel_wait_for_register(&i915->uncore,
> >   					    MIPI_CTRL(port),
> >   					    GLK_PHY_STATUS_PORT_READY, 0, 20))
> >   			DRM_ERROR("PHY is not turning OFF\n");
> > @@ -605,14 +605,14 @@ static void glk_dsi_clear_device_ready(struct intel_encoder *encoder)
> >   static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	DRM_DEBUG_KMS("\n");
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		/* Common bit for both MIPI Port A & MIPI Port C on VLV/CHV */
> > -		i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> > +		i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> >   			BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(PORT_A);
> >   		u32 val;
> > @@ -632,8 +632,8 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> >   		 * On VLV/CHV, wait till Clock lanes are in LP-00 state for MIPI
> >   		 * Port A only. MIPI Port C has no similar bit for checking.
> >   		 */
> > -		if ((IS_GEN9_LP(dev_priv) || port == PORT_A) &&
> > -		    intel_wait_for_register(&dev_priv->uncore,
> > +		if ((IS_GEN9_LP(i915) || port == PORT_A) &&
> > +		    intel_wait_for_register(&i915->uncore,
> >   					    port_ctrl, AFE_LATCHOUT, 0,
> >   					    30))
> >   			DRM_ERROR("DSI LP not going Low\n");
> > @@ -651,14 +651,14 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder)
> >   static void intel_dsi_port_enable(struct intel_encoder *encoder,
> >   				  const struct intel_crtc_state *crtc_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	if (intel_dsi->dual_link == DSI_DUAL_LINK_FRONT_BACK) {
> >   		u32 temp;
> > -		if (IS_GEN9_LP(dev_priv)) {
> > +		if (IS_GEN9_LP(i915)) {
> >   			for_each_dsi_port(port, intel_dsi->ports) {
> >   				temp = I915_READ(MIPI_CTRL(port));
> >   				temp &= ~BXT_PIXEL_OVERLAP_CNT_MASK |
> > @@ -676,7 +676,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> >   	}
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> > +		i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> >   			BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> >   		u32 temp;
> > @@ -688,7 +688,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> >   		if (intel_dsi->ports == (BIT(PORT_A) | BIT(PORT_C))) {
> >   			temp |= (intel_dsi->dual_link - 1)
> >   						<< DUAL_LINK_MODE_SHIFT;
> > -			if (IS_BROXTON(dev_priv))
> > +			if (IS_BROXTON(i915))
> >   				temp |= LANE_CONFIGURATION_DUAL_LINK_A;
> >   			else
> >   				temp |= crtc->pipe ?
> > @@ -708,12 +708,12 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder,
> >   static void intel_dsi_port_disable(struct intel_encoder *encoder)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ?
> > +		i915_reg_t port_ctrl = IS_GEN9_LP(i915) ?
> >   			BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> >   		u32 temp;
> > @@ -775,7 +775,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   {
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	struct drm_crtc *crtc = pipe_config->base.crtc;
> > -	struct drm_i915_private *dev_priv = to_i915(crtc->dev);
> > +	struct drm_i915_private *i915 = to_i915(crtc->dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> >   	int pipe = intel_crtc->pipe;
> >   	enum port port;
> > @@ -784,13 +784,13 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   	DRM_DEBUG_KMS("\n");
> > -	intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true);
> > +	intel_set_cpu_fifo_underrun_reporting(i915, pipe, true);
> >   	/*
> >   	 * The BIOS may leave the PLL in a wonky state where it doesn't
> >   	 * lock. It needs to be fully powered down to fix it.
> >   	 */
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		bxt_dsi_pll_disable(encoder);
> >   		bxt_dsi_pll_enable(encoder, pipe_config);
> >   	} else {
> > @@ -798,7 +798,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   		vlv_dsi_pll_enable(encoder, pipe_config);
> >   	}
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		/* Add MIPI IO reset programming for modeset */
> >   		val = I915_READ(BXT_P_CR_GT_DISP_PWRON);
> >   		I915_WRITE(BXT_P_CR_GT_DISP_PWRON,
> > @@ -809,7 +809,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   		I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, 0);
> >   	}
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		u32 val;
> >   		/* Disable DPOunit clock gating, can stall pipe */
> > @@ -818,7 +818,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   		I915_WRITE(DSPCLK_GATE_D, val);
> >   	}
> > -	if (!IS_GEMINILAKE(dev_priv))
> > +	if (!IS_GEMINILAKE(i915))
> >   		intel_dsi_prepare(encoder, pipe_config);
> >   	/* Power on, try both CRC pmic gpio and VBT */
> > @@ -830,7 +830,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   	/* Deassert reset */
> >   	intel_dsi_vbt_exec_sequence(intel_dsi, MIPI_SEQ_DEASSERT_RESET);
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		glk_cold_boot = glk_dsi_enable_io(encoder);
> >   		/* Prepare port in cold boot(s3/s4) scenario */
> > @@ -842,7 +842,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder,
> >   	intel_dsi_device_ready(encoder);
> >   	/* Prepare port in normal boot scenario */
> > -	if (IS_GEMINILAKE(dev_priv) && !glk_cold_boot)
> > +	if (IS_GEMINILAKE(i915) && !glk_cold_boot)
> >   		intel_dsi_prepare(encoder, pipe_config);
> >   	/* Send initialization commands in LP mode */
> > @@ -901,9 +901,9 @@ static void intel_dsi_disable(struct intel_encoder *encoder,
> >   static void intel_dsi_clear_device_ready(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> > -	if (IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEMINILAKE(i915))
> >   		glk_dsi_clear_device_ready(encoder);
> >   	else
> >   		vlv_dsi_clear_device_ready(encoder);
> > @@ -913,7 +913,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> >   				   const struct intel_crtc_state *pipe_config,
> >   				   const struct drm_connector_state *conn_state)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> > @@ -941,7 +941,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> >   	/* Transition to LP-00 */
> >   	intel_dsi_clear_device_ready(encoder);
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		/* Power down DSI regulator to save power */
> >   		I915_WRITE(BXT_P_DSI_REGULATOR_CFG, STAP_SELECT);
> >   		I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, HS_IO_CTRL_SELECT);
> > @@ -952,7 +952,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> >   				val & ~MIPIO_RST_CTRL);
> >   	}
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		bxt_dsi_pll_disable(encoder);
> >   	} else {
> >   		u32 val;
> > @@ -983,7 +983,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder,
> >   static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   				   enum pipe *pipe)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	intel_wakeref_t wakeref;
> >   	enum port port;
> > @@ -991,7 +991,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   	DRM_DEBUG_KMS("\n");
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv,
> > +	wakeref = intel_display_power_get_if_enabled(i915,
> >   						     encoder->power_domain);
> >   	if (!wakeref)
> >   		return false;
> > @@ -1001,12 +1001,12 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   	 * configuration, otherwise accessing DSI registers will hang the
> >   	 * machine. See BSpec North Display Engine registers/MIPI[BXT].
> >   	 */
> > -	if (IS_GEN9_LP(dev_priv) && !bxt_dsi_pll_is_enabled(dev_priv))
> > +	if (IS_GEN9_LP(i915) && !bxt_dsi_pll_is_enabled(i915))
> >   		goto out_put_power;
> >   	/* XXX: this only works for one DSI output */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		i915_reg_t ctrl_reg = IS_GEN9_LP(dev_priv) ?
> > +		i915_reg_t ctrl_reg = IS_GEN9_LP(i915) ?
> >   			BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port);
> >   		bool enabled = I915_READ(ctrl_reg) & DPI_ENABLE;
> > @@ -1015,7 +1015,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   		 * bit in port C control register does not get set. As a
> >   		 * workaround, check pipe B conf instead.
> >   		 */
> > -		if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > +		if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> >   		    port == PORT_C)
> >   			enabled = I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE;
> > @@ -1031,7 +1031,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   		if (!(I915_READ(MIPI_DEVICE_READY(port)) & DEVICE_READY))
> >   			continue;
> > -		if (IS_GEN9_LP(dev_priv)) {
> > +		if (IS_GEN9_LP(i915)) {
> >   			u32 tmp = I915_READ(MIPI_CTRL(port));
> >   			tmp &= BXT_PIPE_SELECT_MASK;
> >   			tmp >>= BXT_PIPE_SELECT_SHIFT;
> > @@ -1049,7 +1049,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder,
> >   	}
> >   out_put_power:
> > -	intel_display_power_put(dev_priv, encoder->power_domain, wakeref);
> > +	intel_display_power_put(i915, encoder->power_domain, wakeref);
> >   	return active;
> >   }
> > @@ -1058,7 +1058,7 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder,
> >   				    struct intel_crtc_state *pipe_config)
> >   {
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct drm_display_mode *adjusted_mode =
> >   					&pipe_config->base.adjusted_mode;
> >   	struct drm_display_mode *adjusted_mode_sw;
> > @@ -1218,13 +1218,13 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder,
> >   static void intel_dsi_get_config(struct intel_encoder *encoder,
> >   				 struct intel_crtc_state *pipe_config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 pclk;
> >   	DRM_DEBUG_KMS("\n");
> >   	pipe_config->output_types |= BIT(INTEL_OUTPUT_DSI);
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		bxt_dsi_get_pipe_config(encoder, pipe_config);
> >   		pclk = bxt_dsi_get_pclk(encoder, pipe_config);
> >   	} else {
> > @@ -1255,7 +1255,7 @@ static void set_dsi_timings(struct drm_encoder *encoder,
> >   			    const struct drm_display_mode *adjusted_mode)
> >   {
> >   	struct drm_device *dev = encoder->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder);
> >   	enum port port;
> >   	unsigned int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> > @@ -1290,7 +1290,7 @@ static void set_dsi_timings(struct drm_encoder *encoder,
> >   	hbp = txbyteclkhs(hbp, bpp, lane_count, intel_dsi->burst_mode_ratio);
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (IS_GEN9_LP(dev_priv)) {
> > +		if (IS_GEN9_LP(i915)) {
> >   			/*
> >   			 * Program hdisplay and vdisplay on MIPI transcoder.
> >   			 * This is different from calculated hactive and
> > @@ -1342,7 +1342,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   {
> >   	struct drm_encoder *encoder = &intel_encoder->base;
> >   	struct drm_device *dev = encoder->dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder);
> >   	const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode;
> > @@ -1362,7 +1362,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   	}
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > -		if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +		if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   			/*
> >   			 * escape clock divider, 20MHz, shared for A and C.
> >   			 * device ready must be off when doing this! txclkesc?
> > @@ -1377,7 +1377,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   			tmp &= ~READ_REQUEST_PRIORITY_MASK;
> >   			I915_WRITE(MIPI_CTRL(port), tmp |
> >   					READ_REQUEST_PRIORITY_HIGH);
> > -		} else if (IS_GEN9_LP(dev_priv)) {
> > +		} else if (IS_GEN9_LP(i915)) {
> >   			enum pipe pipe = intel_crtc->pipe;
> >   			tmp = I915_READ(MIPI_CTRL(port));
> > @@ -1415,7 +1415,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   	if (intel_dsi->clock_stop)
> >   		tmp |= CLOCKSTOP;
> > -	if (IS_GEN9_LP(dev_priv)) {
> > +	if (IS_GEN9_LP(i915)) {
> >   		tmp |= BXT_DPHY_DEFEATURE_EN;
> >   		if (!is_cmd_mode(intel_dsi))
> >   			tmp |= BXT_DEFEATURE_DPI_FIFO_CTR;
> > @@ -1466,7 +1466,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   		I915_WRITE(MIPI_INIT_COUNT(port),
> >   				txclkesc(intel_dsi->escape_clk_div, 100));
> > -		if (IS_GEN9_LP(dev_priv) && (!intel_dsi->dual_link)) {
> > +		if (IS_GEN9_LP(i915) && (!intel_dsi->dual_link)) {
> >   			/*
> >   			 * BXT spec says write MIPI_INIT_COUNT for
> >   			 * both the ports, even if only one is
> > @@ -1500,7 +1500,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   		 */
> >   		I915_WRITE(MIPI_LP_BYTECLK(port), intel_dsi->lp_byte_clk);
> > -		if (IS_GEMINILAKE(dev_priv)) {
> > +		if (IS_GEMINILAKE(i915)) {
> >   			I915_WRITE(MIPI_TLPX_TIME_COUNT(port),
> >   					intel_dsi->lp_byte_clk);
> >   			/* Shadow of DPHY reg */
> > @@ -1533,19 +1533,19 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder,
> >   static void intel_dsi_unprepare(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> > -	if (IS_GEMINILAKE(dev_priv))
> > +	if (IS_GEMINILAKE(i915))
> >   		return;
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> >   		/* Panel commands can be sent when clock is in LP11 */
> >   		I915_WRITE(MIPI_DEVICE_READY(port), 0x0);
> > -		if (IS_GEN9_LP(dev_priv))
> > +		if (IS_GEN9_LP(i915))
> >   			bxt_dsi_reset_clocks(encoder, port);
> >   		else
> >   			vlv_dsi_reset_clocks(encoder, port);
> > @@ -1594,7 +1594,7 @@ static const struct drm_connector_funcs intel_dsi_connector_funcs = {
> >   static enum drm_panel_orientation
> >   vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	struct intel_encoder *encoder = connector->encoder;
> >   	enum intel_display_power_domain power_domain;
> >   	enum drm_panel_orientation orientation;
> > @@ -1607,11 +1607,11 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> >   	if (!encoder->get_hw_state(encoder, &pipe))
> >   		return DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> > -	crtc = intel_get_crtc_for_pipe(dev_priv, pipe);
> > +	crtc = intel_get_crtc_for_pipe(i915, pipe);
> >   	plane = to_intel_plane(crtc->base.primary);
> >   	power_domain = POWER_DOMAIN_PIPE(pipe);
> > -	wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain);
> > +	wakeref = intel_display_power_get_if_enabled(i915, power_domain);
> >   	if (!wakeref)
> >   		return DRM_MODE_PANEL_ORIENTATION_UNKNOWN;
> > @@ -1624,7 +1624,7 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> >   	else
> >   		orientation = DRM_MODE_PANEL_ORIENTATION_NORMAL;
> > -	intel_display_power_put(dev_priv, power_domain, wakeref);
> > +	intel_display_power_put(i915, power_domain, wakeref);
> >   	return orientation;
> >   }
> > @@ -1632,10 +1632,10 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector)
> >   static enum drm_panel_orientation
> >   vlv_dsi_get_panel_orientation(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	enum drm_panel_orientation orientation;
> > -	if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
> >   		orientation = vlv_dsi_get_hw_panel_orientation(connector);
> >   		if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN)
> >   			return orientation;
> > @@ -1646,13 +1646,13 @@ vlv_dsi_get_panel_orientation(struct intel_connector *connector)
> >   static void intel_dsi_add_properties(struct intel_connector *connector)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(connector->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(connector->base.dev);
> >   	if (connector->panel.fixed_mode) {
> >   		u32 allowed_scalers;
> >   		allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN);
> > -		if (!HAS_GMCH(dev_priv))
> > +		if (!HAS_GMCH(i915))
> >   			allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER);
> >   		drm_connector_attach_scaling_mode_property(&connector->base,
> > @@ -1679,8 +1679,8 @@ static void intel_dsi_add_properties(struct intel_connector *connector)
> >   static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> >   {
> >   	struct drm_device *dev = intel_dsi->base.base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > -	struct mipi_config *mipi_config = dev_priv->vbt.dsi.config;
> > +	struct drm_i915_private *i915 = to_i915(dev);
> > +	struct mipi_config *mipi_config = i915->vbt.dsi.config;
> >   	u32 tlpx_ns, extra_byte_count, tlpx_ui;
> >   	u32 ui_num, ui_den;
> >   	u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt;
> > @@ -1727,7 +1727,7 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> >   	 * For GEMINILAKE dphy_param_reg will be programmed in terms of
> >   	 * HS byte clock count for other platform in HS ddr clock count
> >   	 */
> > -	mul = IS_GEMINILAKE(dev_priv) ? 8 : 2;
> > +	mul = IS_GEMINILAKE(i915) ? 8 : 2;
> >   	ths_prepare_ns = max(mipi_config->ths_prepare,
> >   			     mipi_config->tclk_prepare);
> > @@ -1837,9 +1837,9 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi)
> >   	intel_dsi_log_params(intel_dsi);
> >   }
> > -void vlv_dsi_init(struct drm_i915_private *dev_priv)
> > +void vlv_dsi_init(struct drm_i915_private *i915)
> >   {
> > -	struct drm_device *dev = &dev_priv->drm;
> > +	struct drm_device *dev = &i915->drm;
> >   	struct intel_dsi *intel_dsi;
> >   	struct intel_encoder *intel_encoder;
> >   	struct drm_encoder *encoder;
> > @@ -1851,13 +1851,13 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> >   	DRM_DEBUG_KMS("\n");
> >   	/* There is no detection method for MIPI so rely on VBT */
> > -	if (!intel_bios_is_dsi_present(dev_priv, &port))
> > +	if (!intel_bios_is_dsi_present(i915, &port))
> >   		return;
> > -	if (IS_GEN9_LP(dev_priv))
> > -		dev_priv->mipi_mmio_base = BXT_MIPI_BASE;
> > +	if (IS_GEN9_LP(i915))
> > +		i915->mipi_mmio_base = BXT_MIPI_BASE;
> >   	else
> > -		dev_priv->mipi_mmio_base = VLV_MIPI_BASE;
> > +		i915->mipi_mmio_base = VLV_MIPI_BASE;
> >   	intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL);
> >   	if (!intel_dsi)
> > @@ -1897,20 +1897,20 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> >   	 * On BYT/CHV, pipe A maps to MIPI DSI port A, pipe B maps to MIPI DSI
> >   	 * port C. BXT isn't limited like this.
> >   	 */
> > -	if (IS_GEN9_LP(dev_priv))
> > +	if (IS_GEN9_LP(i915))
> >   		intel_encoder->crtc_mask = BIT(PIPE_A) | BIT(PIPE_B) | BIT(PIPE_C);
> >   	else if (port == PORT_A)
> >   		intel_encoder->crtc_mask = BIT(PIPE_A);
> >   	else
> >   		intel_encoder->crtc_mask = BIT(PIPE_B);
> > -	if (dev_priv->vbt.dsi.config->dual_link)
> > +	if (i915->vbt.dsi.config->dual_link)
> >   		intel_dsi->ports = BIT(PORT_A) | BIT(PORT_C);
> >   	else
> >   		intel_dsi->ports = BIT(port);
> > -	intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports;
> > -	intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports;
> > +	intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports;
> > +	intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports;
> >   	/* Create a DSI host (and a device) for each port. */
> >   	for_each_dsi_port(port, intel_dsi->ports) {
> > @@ -1949,8 +1949,8 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv)
> >   	 * In case of BYT with CRC PMIC, we need to use GPIO for
> >   	 * Panel control.
> >   	 */
> > -	if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) &&
> > -	    (dev_priv->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) {
> > +	if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) &&
> > +	    (i915->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) {
> >   		intel_dsi->gpio_panel =
> >   			gpiod_get(dev->dev, "panel", GPIOD_OUT_HIGH);
> > diff --git a/drivers/gpu/drm/i915/vlv_dsi_pll.c b/drivers/gpu/drm/i915/vlv_dsi_pll.c
> > index 99cc3e2e9c2c..26655d5472b4 100644
> > --- a/drivers/gpu/drm/i915/vlv_dsi_pll.c
> > +++ b/drivers/gpu/drm/i915/vlv_dsi_pll.c
> > @@ -53,7 +53,7 @@ static u32 dsi_clk_from_pclk(u32 pclk, enum mipi_dsi_pixel_format fmt,
> >   	return dsi_clk_khz;
> >   }
> > -static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> > +static int dsi_calc_mnp(struct drm_i915_private *i915,
> >   			struct intel_crtc_state *config,
> >   			int target_dsi_clk)
> >   {
> > @@ -68,7 +68,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> >   		return -ECHRNG;
> >   	}
> > -	if (IS_CHERRYVIEW(dev_priv)) {
> > +	if (IS_CHERRYVIEW(i915)) {
> >   		ref_clk = 100000;
> >   		n = 4;
> >   		m_min = 70;
> > @@ -116,7 +116,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv,
> >   int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> >   			struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	int ret;
> >   	u32 dsi_clk;
> > @@ -124,7 +124,7 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> >   	dsi_clk = dsi_clk_from_pclk(intel_dsi->pclk, intel_dsi->pixel_format,
> >   				    intel_dsi->lane_count);
> > -	ret = dsi_calc_mnp(dev_priv, config, dsi_clk);
> > +	ret = dsi_calc_mnp(i915, config, dsi_clk);
> >   	if (ret) {
> >   		DRM_DEBUG_KMS("dsi_calc_mnp failed\n");
> >   		return ret;
> > @@ -147,15 +147,15 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder,
> >   void vlv_dsi_pll_enable(struct intel_encoder *encoder,
> >   			const struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	DRM_DEBUG_KMS("\n");
> > -	vlv_cck_get(dev_priv);
> > +	vlv_cck_get(i915);
> > -	vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, 0);
> > -	vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div);
> > -	vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL,
> > +	vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, 0);
> > +	vlv_cck_write(i915, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div);
> > +	vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL,
> >   		      config->dsi_pll.ctrl & ~DSI_PLL_VCO_EN);
> >   	/* wait at least 0.5 us after ungating before enabling VCO,
> > @@ -163,38 +163,38 @@ void vlv_dsi_pll_enable(struct intel_encoder *encoder,
> >   	 */
> >   	usleep_range(10, 50);
> > -	vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl);
> > +	vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl);
> > -	if (wait_for(vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL) &
> > +	if (wait_for(vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL) &
> >   						DSI_PLL_LOCK, 20)) {
> > -		vlv_cck_put(dev_priv);
> > +		vlv_cck_put(i915);
> >   		DRM_ERROR("DSI PLL lock failed\n");
> >   		return;
> >   	}
> > -	vlv_cck_put(dev_priv);
> > +	vlv_cck_put(i915);
> >   	DRM_DEBUG_KMS("DSI PLL locked\n");
> >   }
> >   void vlv_dsi_pll_disable(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 tmp;
> >   	DRM_DEBUG_KMS("\n");
> > -	vlv_cck_get(dev_priv);
> > +	vlv_cck_get(i915);
> > -	tmp = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> > +	tmp = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> >   	tmp &= ~DSI_PLL_VCO_EN;
> >   	tmp |= DSI_PLL_LDO_GATE;
> > -	vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, tmp);
> > +	vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, tmp);
> > -	vlv_cck_put(dev_priv);
> > +	vlv_cck_put(i915);
> >   }
> > -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
> > +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915)
> >   {
> >   	bool enabled;
> >   	u32 val;
> > @@ -216,7 +216,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
> >   	 * causes a system hang.
> >   	 */
> >   	val = I915_READ(BXT_DSI_PLL_CTL);
> > -	if (IS_GEMINILAKE(dev_priv)) {
> > +	if (IS_GEMINILAKE(i915)) {
> >   		if (!(val & BXT_DSIA_16X_MASK)) {
> >   			DRM_DEBUG_DRIVER("Invalid PLL divider (%08x)\n", val);
> >   			enabled = false;
> > @@ -233,7 +233,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv)
> >   void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	u32 val;
> >   	DRM_DEBUG_KMS("\n");
> > @@ -246,7 +246,7 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> >   	 * PLL lock should deassert within 200us.
> >   	 * Wait up to 1ms before timing out.
> >   	 */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    BXT_DSI_PLL_ENABLE,
> >   				    BXT_DSI_PLL_LOCKED,
> >   				    0,
> > @@ -257,21 +257,21 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder)
> >   u32 vlv_dsi_get_pclk(struct intel_encoder *encoder,
> >   		     struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> >   	u32 dsi_clock, pclk;
> >   	u32 pll_ctl, pll_div;
> >   	u32 m = 0, p = 0, n;
> > -	int refclk = IS_CHERRYVIEW(dev_priv) ? 100000 : 25000;
> > +	int refclk = IS_CHERRYVIEW(i915) ? 100000 : 25000;
> >   	int i;
> >   	DRM_DEBUG_KMS("\n");
> > -	vlv_cck_get(dev_priv);
> > -	pll_ctl = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL);
> > -	pll_div = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_DIVIDER);
> > -	vlv_cck_put(dev_priv);
> > +	vlv_cck_get(i915);
> > +	pll_ctl = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL);
> > +	pll_div = vlv_cck_read(i915, CCK_REG_DSI_PLL_DIVIDER);
> > +	vlv_cck_put(i915);
> >   	config->dsi_pll.ctrl = pll_ctl & ~DSI_PLL_LOCK;
> >   	config->dsi_pll.div = pll_div;
> > @@ -325,7 +325,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder,
> >   	u32 dsi_clk;
> >   	u32 dsi_ratio;
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format);
> >   	config->dsi_pll.ctrl = I915_READ(BXT_DSI_PLL_CTL);
> > @@ -343,7 +343,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder,
> >   void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> >   {
> >   	u32 temp;
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	temp = I915_READ(MIPI_CTRL(port));
> > @@ -356,7 +356,7 @@ void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> >   static void glk_dsi_program_esc_clock(struct drm_device *dev,
> >   				   const struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 dsi_rate = 0;
> >   	u32 pll_ratio = 0;
> >   	u32 ddr_clk = 0;
> > @@ -404,7 +404,7 @@ static void glk_dsi_program_esc_clock(struct drm_device *dev,
> >   static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port,
> >   				   const struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	u32 tmp;
> >   	u32 dsi_rate = 0;
> >   	u32 pll_ratio = 0;
> > @@ -457,7 +457,7 @@ static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port,
> >   int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> >   			struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	u8 dsi_ratio, dsi_ratio_min, dsi_ratio_max;
> >   	u32 dsi_clk;
> > @@ -472,7 +472,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> >   	 */
> >   	dsi_ratio = DIV_ROUND_UP(dsi_clk * 2, BXT_REF_CLOCK_KHZ);
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		dsi_ratio_min = BXT_DSI_PLL_RATIO_MIN;
> >   		dsi_ratio_max = BXT_DSI_PLL_RATIO_MAX;
> >   	} else {
> > @@ -496,7 +496,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> >   	/* As per recommendation from hardware team,
> >   	 * Prog PVD ratio =1 if dsi ratio <= 50
> >   	 */
> > -	if (IS_BROXTON(dev_priv) && dsi_ratio <= 50)
> > +	if (IS_BROXTON(i915) && dsi_ratio <= 50)
> >   		config->dsi_pll.ctrl |= BXT_DSI_PLL_PVD_RATIO_1;
> >   	return 0;
> > @@ -505,7 +505,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder,
> >   void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> >   			const struct intel_crtc_state *config)
> >   {
> > -	struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
> > +	struct drm_i915_private *i915 = to_i915(encoder->base.dev);
> >   	struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base);
> >   	enum port port;
> >   	u32 val;
> > @@ -517,7 +517,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> >   	POSTING_READ(BXT_DSI_PLL_CTL);
> >   	/* Program TX, RX, Dphy clocks */
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		for_each_dsi_port(port, intel_dsi->ports)
> >   			bxt_dsi_program_clocks(encoder->base.dev, port, config);
> >   	} else {
> > @@ -530,7 +530,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder,
> >   	I915_WRITE(BXT_DSI_PLL_ENABLE, val);
> >   	/* Timeout and fail if PLL not locked */
> > -	if (intel_wait_for_register(&dev_priv->uncore,
> > +	if (intel_wait_for_register(&i915->uncore,
> >   				    BXT_DSI_PLL_ENABLE,
> >   				    BXT_DSI_PLL_LOCKED,
> >   				    BXT_DSI_PLL_LOCKED,
> > @@ -546,10 +546,10 @@ void bxt_dsi_reset_clocks(struct intel_encoder *encoder, enum port port)
> >   {
> >   	u32 tmp;
> >   	struct drm_device *dev = encoder->base.dev;
> > -	struct drm_i915_private *dev_priv = to_i915(dev);
> > +	struct drm_i915_private *i915 = to_i915(dev);
> >   	/* Clear old configurations */
> > -	if (IS_BROXTON(dev_priv)) {
> > +	if (IS_BROXTON(i915)) {
> >   		tmp = I915_READ(BXT_MIPI_CLOCK_CTL);
> >   		tmp &= ~(BXT_MIPI_TX_ESCLK_FIXDIV_MASK(port));
> >   		tmp &= ~(BXT_MIPI_RX_ESCLK_UPPER_FIXDIV_MASK(port));
> > 


More information about the Intel-gfx mailing list