[Intel-gfx] [PATCH 06/50] drm/i915: s/intel_ring_buffer/intel_engine

Daniel Vetter daniel at ffwll.ch
Tue May 13 15:28:27 CEST 2014


On Fri, May 09, 2014 at 01:08:36PM +0100, oscar.mateo at intel.com wrote:
> From: Oscar Mateo <oscar.mateo at intel.com>
> 
> In the upcoming patches, we plan to break the correlation between
> engines (a.k.a. rings) and ringbuffers, so it makes sense to
> refactor the code and make the change obvious.
> 
> No functional changes.
> 
> Signed-off-by: Oscar Mateo <oscar.mateo at intel.com>

If we rename stuff I'd vote for something close to Bspec language, like
CS. So maybe intel_cs_engine?

/me sucks at this naming game
-Daniel

> ---
>  drivers/gpu/drm/i915/i915_cmd_parser.c     |  16 +--
>  drivers/gpu/drm/i915/i915_debugfs.c        |  16 +--
>  drivers/gpu/drm/i915/i915_dma.c            |  10 +-
>  drivers/gpu/drm/i915/i915_drv.h            |  32 +++---
>  drivers/gpu/drm/i915/i915_gem.c            |  58 +++++------
>  drivers/gpu/drm/i915/i915_gem_context.c    |  14 +--
>  drivers/gpu/drm/i915/i915_gem_execbuffer.c |  18 ++--
>  drivers/gpu/drm/i915/i915_gem_gtt.c        |  18 ++--
>  drivers/gpu/drm/i915/i915_gem_gtt.h        |   2 +-
>  drivers/gpu/drm/i915/i915_gpu_error.c      |   6 +-
>  drivers/gpu/drm/i915/i915_irq.c            |  28 ++---
>  drivers/gpu/drm/i915/i915_trace.h          |  26 ++---
>  drivers/gpu/drm/i915/intel_display.c       |  14 +--
>  drivers/gpu/drm/i915/intel_drv.h           |   4 +-
>  drivers/gpu/drm/i915/intel_overlay.c       |  12 +--
>  drivers/gpu/drm/i915/intel_pm.c            |  10 +-
>  drivers/gpu/drm/i915/intel_ringbuffer.c    | 158 ++++++++++++++---------------
>  drivers/gpu/drm/i915/intel_ringbuffer.h    |  76 +++++++-------
>  18 files changed, 259 insertions(+), 259 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c
> index 69d34e4..3234d36 100644
> --- a/drivers/gpu/drm/i915/i915_cmd_parser.c
> +++ b/drivers/gpu/drm/i915/i915_cmd_parser.c
> @@ -498,7 +498,7 @@ static u32 gen7_blt_get_cmd_length_mask(u32 cmd_header)
>  	return 0;
>  }
>  
> -static bool validate_cmds_sorted(struct intel_ring_buffer *ring)
> +static bool validate_cmds_sorted(struct intel_engine *ring)
>  {
>  	int i;
>  	bool ret = true;
> @@ -550,7 +550,7 @@ static bool check_sorted(int ring_id, const u32 *reg_table, int reg_count)
>  	return ret;
>  }
>  
> -static bool validate_regs_sorted(struct intel_ring_buffer *ring)
> +static bool validate_regs_sorted(struct intel_engine *ring)
>  {
>  	return check_sorted(ring->id, ring->reg_table, ring->reg_count) &&
>  		check_sorted(ring->id, ring->master_reg_table,
> @@ -562,10 +562,10 @@ static bool validate_regs_sorted(struct intel_ring_buffer *ring)
>   * @ring: the ringbuffer to initialize
>   *
>   * Optionally initializes fields related to batch buffer command parsing in the
> - * struct intel_ring_buffer based on whether the platform requires software
> + * struct intel_engine based on whether the platform requires software
>   * command parsing.
>   */
> -void i915_cmd_parser_init_ring(struct intel_ring_buffer *ring)
> +void i915_cmd_parser_init_ring(struct intel_engine *ring)
>  {
>  	if (!IS_GEN7(ring->dev))
>  		return;
> @@ -664,7 +664,7 @@ find_cmd_in_table(const struct drm_i915_cmd_table *table,
>   * ring's default length encoding and returns default_desc.
>   */
>  static const struct drm_i915_cmd_descriptor*
> -find_cmd(struct intel_ring_buffer *ring,
> +find_cmd(struct intel_engine *ring,
>  	 u32 cmd_header,
>  	 struct drm_i915_cmd_descriptor *default_desc)
>  {
> @@ -744,7 +744,7 @@ finish:
>   *
>   * Return: true if the ring requires software command parsing
>   */
> -bool i915_needs_cmd_parser(struct intel_ring_buffer *ring)
> +bool i915_needs_cmd_parser(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  
> @@ -763,7 +763,7 @@ bool i915_needs_cmd_parser(struct intel_ring_buffer *ring)
>  	return (i915.enable_cmd_parser == 1);
>  }
>  
> -static bool check_cmd(const struct intel_ring_buffer *ring,
> +static bool check_cmd(const struct intel_engine *ring,
>  		      const struct drm_i915_cmd_descriptor *desc,
>  		      const u32 *cmd,
>  		      const bool is_master,
> @@ -865,7 +865,7 @@ static bool check_cmd(const struct intel_ring_buffer *ring,
>   *
>   * Return: non-zero if the parser finds violations or otherwise fails
>   */
> -int i915_parse_cmds(struct intel_ring_buffer *ring,
> +int i915_parse_cmds(struct intel_engine *ring,
>  		    struct drm_i915_gem_object *batch_obj,
>  		    u32 batch_start_offset,
>  		    bool is_master)
> diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
> index 103e62c..0052460 100644
> --- a/drivers/gpu/drm/i915/i915_debugfs.c
> +++ b/drivers/gpu/drm/i915/i915_debugfs.c
> @@ -562,7 +562,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	struct drm_i915_gem_request *gem_request;
>  	int ret, count, i;
>  
> @@ -594,7 +594,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data)
>  }
>  
>  static void i915_ring_seqno_info(struct seq_file *m,
> -				 struct intel_ring_buffer *ring)
> +				 struct intel_engine *ring)
>  {
>  	if (ring->get_seqno) {
>  		seq_printf(m, "Current sequence (%s): %u\n",
> @@ -607,7 +607,7 @@ static int i915_gem_seqno_info(struct seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int ret, i;
>  
>  	ret = mutex_lock_interruptible(&dev->struct_mutex);
> @@ -630,7 +630,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int ret, i, pipe;
>  
>  	ret = mutex_lock_interruptible(&dev->struct_mutex);
> @@ -800,7 +800,7 @@ static int i915_hws_info(struct seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	const u32 *hws;
>  	int i;
>  
> @@ -1677,7 +1677,7 @@ static int i915_context_status(struct seq_file *m, void *unused)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	struct i915_hw_context *ctx;
>  	int ret, i;
>  
> @@ -1826,7 +1826,7 @@ static int per_file_ctx(int id, void *ptr, void *data)
>  static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	struct i915_hw_ppgtt *ppgtt = dev_priv->mm.aliasing_ppgtt;
>  	int unused, i;
>  
> @@ -1850,7 +1850,7 @@ static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev)
>  static void gen6_ppgtt_info(struct seq_file *m, struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	struct drm_file *file;
>  	int i;
>  
> diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c
> index d02c8de..5263d63 100644
> --- a/drivers/gpu/drm/i915/i915_dma.c
> +++ b/drivers/gpu/drm/i915/i915_dma.c
> @@ -119,7 +119,7 @@ static void i915_write_hws_pga(struct drm_device *dev)
>  static void i915_free_hws(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = LP_RING(dev_priv);
> +	struct intel_engine *ring = LP_RING(dev_priv);
>  
>  	if (dev_priv->status_page_dmah) {
>  		drm_pci_free(dev, dev_priv->status_page_dmah);
> @@ -139,7 +139,7 @@ void i915_kernel_lost_context(struct drm_device * dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct drm_i915_master_private *master_priv;
> -	struct intel_ring_buffer *ring = LP_RING(dev_priv);
> +	struct intel_engine *ring = LP_RING(dev_priv);
>  
>  	/*
>  	 * We should never lose context on the ring with modesetting
> @@ -234,7 +234,7 @@ static int i915_initialize(struct drm_device * dev, drm_i915_init_t * init)
>  static int i915_dma_resume(struct drm_device * dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = LP_RING(dev_priv);
> +	struct intel_engine *ring = LP_RING(dev_priv);
>  
>  	DRM_DEBUG_DRIVER("%s\n", __func__);
>  
> @@ -782,7 +782,7 @@ static int i915_wait_irq(struct drm_device * dev, int irq_nr)
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
>  	int ret = 0;
> -	struct intel_ring_buffer *ring = LP_RING(dev_priv);
> +	struct intel_engine *ring = LP_RING(dev_priv);
>  
>  	DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr,
>  		  READ_BREADCRUMB(dev_priv));
> @@ -1073,7 +1073,7 @@ static int i915_set_status_page(struct drm_device *dev, void *data,
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	drm_i915_hws_addr_t *hws = data;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  
>  	if (drm_core_check_feature(dev, DRIVER_MODESET))
>  		return -ENODEV;
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index b1725c6..3b7a36f9 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -594,7 +594,7 @@ struct i915_hw_context {
>  	bool is_initialized;
>  	uint8_t remap_slice;
>  	struct drm_i915_file_private *file_priv;
> -	struct intel_ring_buffer *last_ring;
> +	struct intel_engine *last_ring;
>  	struct drm_i915_gem_object *obj;
>  	struct i915_ctx_hang_stats hang_stats;
>  	struct i915_address_space *vm;
> @@ -1354,7 +1354,7 @@ struct drm_i915_private {
>  	wait_queue_head_t gmbus_wait_queue;
>  
>  	struct pci_dev *bridge_dev;
> -	struct intel_ring_buffer ring[I915_NUM_RINGS];
> +	struct intel_engine ring[I915_NUM_RINGS];
>  	uint32_t last_seqno, next_seqno;
>  
>  	drm_dma_handle_t *status_page_dmah;
> @@ -1675,7 +1675,7 @@ struct drm_i915_gem_object {
>  	void *dma_buf_vmapping;
>  	int vmapping_count;
>  
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  
>  	/** Breadcrumb of last rendering to the buffer. */
>  	uint32_t last_read_seqno;
> @@ -1714,7 +1714,7 @@ struct drm_i915_gem_object {
>   */
>  struct drm_i915_gem_request {
>  	/** On Which ring this request was generated */
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  
>  	/** GEM sequence number associated with this request. */
>  	uint32_t seqno;
> @@ -1755,7 +1755,7 @@ struct drm_i915_file_private {
>  
>  	struct i915_hw_context *private_default_ctx;
>  	atomic_t rps_wait_boost;
> -	struct  intel_ring_buffer *bsd_ring;
> +	struct  intel_engine *bsd_ring;
>  };
>  
>  /*
> @@ -2182,9 +2182,9 @@ static inline void i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj)
>  
>  int __must_check i915_mutex_lock_interruptible(struct drm_device *dev);
>  int i915_gem_object_sync(struct drm_i915_gem_object *obj,
> -			 struct intel_ring_buffer *to);
> +			 struct intel_engine *to);
>  void i915_vma_move_to_active(struct i915_vma *vma,
> -			     struct intel_ring_buffer *ring);
> +			     struct intel_engine *ring);
>  int i915_gem_dumb_create(struct drm_file *file_priv,
>  			 struct drm_device *dev,
>  			 struct drm_mode_create_dumb *args);
> @@ -2226,7 +2226,7 @@ i915_gem_object_unpin_fence(struct drm_i915_gem_object *obj)
>  }
>  
>  struct drm_i915_gem_request *
> -i915_gem_find_active_request(struct intel_ring_buffer *ring);
> +i915_gem_find_active_request(struct intel_engine *ring);
>  
>  bool i915_gem_retire_requests(struct drm_device *dev);
>  int __must_check i915_gem_check_wedge(struct i915_gpu_error *error,
> @@ -2264,18 +2264,18 @@ bool i915_gem_clflush_object(struct drm_i915_gem_object *obj, bool force);
>  int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj);
>  int __must_check i915_gem_init(struct drm_device *dev);
>  int __must_check i915_gem_init_hw(struct drm_device *dev);
> -int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice);
> +int i915_gem_l3_remap(struct intel_engine *ring, int slice);
>  void i915_gem_init_swizzling(struct drm_device *dev);
>  void i915_gem_cleanup_ringbuffer(struct drm_device *dev);
>  int __must_check i915_gpu_idle(struct drm_device *dev);
>  int __must_check i915_gem_suspend(struct drm_device *dev);
> -int __i915_add_request(struct intel_ring_buffer *ring,
> +int __i915_add_request(struct intel_engine *ring,
>  		       struct drm_file *file,
>  		       struct drm_i915_gem_object *batch_obj,
>  		       u32 *seqno);
>  #define i915_add_request(ring, seqno) \
>  	__i915_add_request(ring, NULL, NULL, seqno)
> -int __must_check i915_wait_seqno(struct intel_ring_buffer *ring,
> +int __must_check i915_wait_seqno(struct intel_engine *ring,
>  				 uint32_t seqno);
>  int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
>  int __must_check
> @@ -2286,7 +2286,7 @@ i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write);
>  int __must_check
>  i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
>  				     u32 alignment,
> -				     struct intel_ring_buffer *pipelined);
> +				     struct intel_engine *pipelined);
>  void i915_gem_object_unpin_from_display_plane(struct drm_i915_gem_object *obj);
>  int i915_gem_attach_phys_object(struct drm_device *dev,
>  				struct drm_i915_gem_object *obj,
> @@ -2388,7 +2388,7 @@ void i915_gem_context_reset(struct drm_device *dev);
>  int i915_gem_context_open(struct drm_device *dev, struct drm_file *file);
>  int i915_gem_context_enable(struct drm_i915_private *dev_priv);
>  void i915_gem_context_close(struct drm_device *dev, struct drm_file *file);
> -int i915_switch_context(struct intel_ring_buffer *ring,
> +int i915_switch_context(struct intel_engine *ring,
>  			struct i915_hw_context *to);
>  struct i915_hw_context *
>  i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id);
> @@ -2497,9 +2497,9 @@ const char *i915_cache_level_str(int type);
>  
>  /* i915_cmd_parser.c */
>  int i915_cmd_parser_get_version(void);
> -void i915_cmd_parser_init_ring(struct intel_ring_buffer *ring);
> -bool i915_needs_cmd_parser(struct intel_ring_buffer *ring);
> -int i915_parse_cmds(struct intel_ring_buffer *ring,
> +void i915_cmd_parser_init_ring(struct intel_engine *ring);
> +bool i915_needs_cmd_parser(struct intel_engine *ring);
> +int i915_parse_cmds(struct intel_engine *ring,
>  		    struct drm_i915_gem_object *batch_obj,
>  		    u32 batch_start_offset,
>  		    bool is_master);
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 6ef53bd..a3b697b 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -64,7 +64,7 @@ static unsigned long i915_gem_inactive_scan(struct shrinker *shrinker,
>  static unsigned long i915_gem_purge(struct drm_i915_private *dev_priv, long target);
>  static unsigned long i915_gem_shrink_all(struct drm_i915_private *dev_priv);
>  static void i915_gem_object_truncate(struct drm_i915_gem_object *obj);
> -static void i915_gem_retire_requests_ring(struct intel_ring_buffer *ring);
> +static void i915_gem_retire_requests_ring(struct intel_engine *ring);
>  
>  static bool cpu_cache_is_coherent(struct drm_device *dev,
>  				  enum i915_cache_level level)
> @@ -977,7 +977,7 @@ i915_gem_check_wedge(struct i915_gpu_error *error,
>   * equal.
>   */
>  static int
> -i915_gem_check_olr(struct intel_ring_buffer *ring, u32 seqno)
> +i915_gem_check_olr(struct intel_engine *ring, u32 seqno)
>  {
>  	int ret;
>  
> @@ -996,7 +996,7 @@ static void fake_irq(unsigned long data)
>  }
>  
>  static bool missed_irq(struct drm_i915_private *dev_priv,
> -		       struct intel_ring_buffer *ring)
> +		       struct intel_engine *ring)
>  {
>  	return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings);
>  }
> @@ -1027,7 +1027,7 @@ static bool can_wait_boost(struct drm_i915_file_private *file_priv)
>   * Returns 0 if the seqno was found within the alloted time. Else returns the
>   * errno with remaining time filled in timeout argument.
>   */
> -static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno,
> +static int __wait_seqno(struct intel_engine *ring, u32 seqno,
>  			unsigned reset_counter,
>  			bool interruptible,
>  			struct timespec *timeout,
> @@ -1134,7 +1134,7 @@ static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno,
>   * request and object lists appropriately for that event.
>   */
>  int
> -i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno)
> +i915_wait_seqno(struct intel_engine *ring, uint32_t seqno)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1159,7 +1159,7 @@ i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno)
>  
>  static int
>  i915_gem_object_wait_rendering__tail(struct drm_i915_gem_object *obj,
> -				     struct intel_ring_buffer *ring)
> +				     struct intel_engine *ring)
>  {
>  	if (!obj->active)
>  		return 0;
> @@ -1184,7 +1184,7 @@ static __must_check int
>  i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj,
>  			       bool readonly)
>  {
> -	struct intel_ring_buffer *ring = obj->ring;
> +	struct intel_engine *ring = obj->ring;
>  	u32 seqno;
>  	int ret;
>  
> @@ -1209,7 +1209,7 @@ i915_gem_object_wait_rendering__nonblocking(struct drm_i915_gem_object *obj,
>  {
>  	struct drm_device *dev = obj->base.dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = obj->ring;
> +	struct intel_engine *ring = obj->ring;
>  	unsigned reset_counter;
>  	u32 seqno;
>  	int ret;
> @@ -2011,7 +2011,7 @@ i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
>  
>  static void
>  i915_gem_object_move_to_active(struct drm_i915_gem_object *obj,
> -			       struct intel_ring_buffer *ring)
> +			       struct intel_engine *ring)
>  {
>  	struct drm_device *dev = obj->base.dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -2049,7 +2049,7 @@ i915_gem_object_move_to_active(struct drm_i915_gem_object *obj,
>  }
>  
>  void i915_vma_move_to_active(struct i915_vma *vma,
> -			     struct intel_ring_buffer *ring)
> +			     struct intel_engine *ring)
>  {
>  	list_move_tail(&vma->mm_list, &vma->vm->active_list);
>  	return i915_gem_object_move_to_active(vma->obj, ring);
> @@ -2090,7 +2090,7 @@ i915_gem_object_move_to_inactive(struct drm_i915_gem_object *obj)
>  static void
>  i915_gem_object_retire(struct drm_i915_gem_object *obj)
>  {
> -	struct intel_ring_buffer *ring = obj->ring;
> +	struct intel_engine *ring = obj->ring;
>  
>  	if (ring == NULL)
>  		return;
> @@ -2104,7 +2104,7 @@ static int
>  i915_gem_init_seqno(struct drm_device *dev, u32 seqno)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int ret, i, j;
>  
>  	/* Carefully retire all requests without writing to the rings */
> @@ -2170,7 +2170,7 @@ i915_gem_get_seqno(struct drm_device *dev, u32 *seqno)
>  	return 0;
>  }
>  
> -int __i915_add_request(struct intel_ring_buffer *ring,
> +int __i915_add_request(struct intel_engine *ring,
>  		       struct drm_file *file,
>  		       struct drm_i915_gem_object *obj,
>  		       u32 *out_seqno)
> @@ -2330,7 +2330,7 @@ static void i915_gem_free_request(struct drm_i915_gem_request *request)
>  }
>  
>  struct drm_i915_gem_request *
> -i915_gem_find_active_request(struct intel_ring_buffer *ring)
> +i915_gem_find_active_request(struct intel_engine *ring)
>  {
>  	struct drm_i915_gem_request *request;
>  	u32 completed_seqno;
> @@ -2348,7 +2348,7 @@ i915_gem_find_active_request(struct intel_ring_buffer *ring)
>  }
>  
>  static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
> -				       struct intel_ring_buffer *ring)
> +				       struct intel_engine *ring)
>  {
>  	struct drm_i915_gem_request *request;
>  	bool ring_hung;
> @@ -2367,7 +2367,7 @@ static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
>  }
>  
>  static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv,
> -					struct intel_ring_buffer *ring)
> +					struct intel_engine *ring)
>  {
>  	while (!list_empty(&ring->active_list)) {
>  		struct drm_i915_gem_object *obj;
> @@ -2426,7 +2426,7 @@ void i915_gem_restore_fences(struct drm_device *dev)
>  void i915_gem_reset(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	/*
> @@ -2449,7 +2449,7 @@ void i915_gem_reset(struct drm_device *dev)
>   * This function clears the request list as sequence numbers are passed.
>   */
>  static void
> -i915_gem_retire_requests_ring(struct intel_ring_buffer *ring)
> +i915_gem_retire_requests_ring(struct intel_engine *ring)
>  {
>  	uint32_t seqno;
>  
> @@ -2512,7 +2512,7 @@ bool
>  i915_gem_retire_requests(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	bool idle = true;
>  	int i;
>  
> @@ -2606,7 +2606,7 @@ i915_gem_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct drm_i915_gem_wait *args = data;
>  	struct drm_i915_gem_object *obj;
> -	struct intel_ring_buffer *ring = NULL;
> +	struct intel_engine *ring = NULL;
>  	struct timespec timeout_stack, *timeout = NULL;
>  	unsigned reset_counter;
>  	u32 seqno = 0;
> @@ -2677,9 +2677,9 @@ out:
>   */
>  int
>  i915_gem_object_sync(struct drm_i915_gem_object *obj,
> -		     struct intel_ring_buffer *to)
> +		     struct intel_engine *to)
>  {
> -	struct intel_ring_buffer *from = obj->ring;
> +	struct intel_engine *from = obj->ring;
>  	u32 seqno;
>  	int ret, idx;
>  
> @@ -2800,7 +2800,7 @@ int i915_vma_unbind(struct i915_vma *vma)
>  int i915_gpu_idle(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int ret, i;
>  
>  	/* Flush everything onto the inactive list. */
> @@ -3659,7 +3659,7 @@ static bool is_pin_display(struct drm_i915_gem_object *obj)
>  int
>  i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
>  				     u32 alignment,
> -				     struct intel_ring_buffer *pipelined)
> +				     struct intel_engine *pipelined)
>  {
>  	u32 old_read_domains, old_write_domain;
>  	int ret;
> @@ -3812,7 +3812,7 @@ i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file)
>  	struct drm_i915_file_private *file_priv = file->driver_priv;
>  	unsigned long recent_enough = jiffies - msecs_to_jiffies(20);
>  	struct drm_i915_gem_request *request;
> -	struct intel_ring_buffer *ring = NULL;
> +	struct intel_engine *ring = NULL;
>  	unsigned reset_counter;
>  	u32 seqno = 0;
>  	int ret;
> @@ -4258,7 +4258,7 @@ static void
>  i915_gem_stop_ringbuffers(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	for_each_active_ring(ring, dev_priv, i)
> @@ -4307,7 +4307,7 @@ err:
>  	return ret;
>  }
>  
> -int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice)
> +int i915_gem_l3_remap(struct intel_engine *ring, int slice)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -4532,7 +4532,7 @@ void
>  i915_gem_cleanup_ringbuffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	for_each_active_ring(ring, dev_priv, i)
> @@ -4608,7 +4608,7 @@ i915_gem_lastclose(struct drm_device *dev)
>  }
>  
>  static void
> -init_ring_lists(struct intel_ring_buffer *ring)
> +init_ring_lists(struct intel_engine *ring)
>  {
>  	INIT_LIST_HEAD(&ring->active_list);
>  	INIT_LIST_HEAD(&ring->request_list);
> diff --git a/drivers/gpu/drm/i915/i915_gem_context.c b/drivers/gpu/drm/i915/i915_gem_context.c
> index 014fb8f..4d37e20 100644
> --- a/drivers/gpu/drm/i915/i915_gem_context.c
> +++ b/drivers/gpu/drm/i915/i915_gem_context.c
> @@ -359,7 +359,7 @@ err_destroy:
>  void i915_gem_context_reset(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	/* Prevent the hardware from restoring the last context (which hung) on
> @@ -392,7 +392,7 @@ int i915_gem_context_init(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct i915_hw_context *ctx;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int unused;
>  
>  	/* Init should only be called once per module load. Eventually the
> @@ -428,7 +428,7 @@ void i915_gem_context_fini(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct i915_hw_context *dctx = dev_priv->ring[RCS].default_context;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int unused;
>  
>  	if (dctx->obj) {
> @@ -467,7 +467,7 @@ void i915_gem_context_fini(struct drm_device *dev)
>  
>  int i915_gem_context_enable(struct drm_i915_private *dev_priv)
>  {
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int ret, i;
>  
>  	/* This is the only place the aliasing PPGTT gets enabled, which means
> @@ -546,7 +546,7 @@ i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id)
>  }
>  
>  static inline int
> -mi_set_context(struct intel_ring_buffer *ring,
> +mi_set_context(struct intel_engine *ring,
>  	       struct i915_hw_context *new_context,
>  	       u32 hw_flags)
>  {
> @@ -596,7 +596,7 @@ mi_set_context(struct intel_ring_buffer *ring,
>  	return ret;
>  }
>  
> -static int do_switch(struct intel_ring_buffer *ring,
> +static int do_switch(struct intel_engine *ring,
>  		     struct i915_hw_context *to)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> @@ -726,7 +726,7 @@ unpin_out:
>   * it will have a refoucnt > 1. This allows us to destroy the context abstract
>   * object while letting the normal object tracking destroy the backing BO.
>   */
> -int i915_switch_context(struct intel_ring_buffer *ring,
> +int i915_switch_context(struct intel_engine *ring,
>  			struct i915_hw_context *to)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> index 47fe8ec..95e797e 100644
> --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
> @@ -541,7 +541,7 @@ need_reloc_mappable(struct i915_vma *vma)
>  
>  static int
>  i915_gem_execbuffer_reserve_vma(struct i915_vma *vma,
> -				struct intel_ring_buffer *ring,
> +				struct intel_engine *ring,
>  				bool *need_reloc)
>  {
>  	struct drm_i915_gem_object *obj = vma->obj;
> @@ -596,7 +596,7 @@ i915_gem_execbuffer_reserve_vma(struct i915_vma *vma,
>  }
>  
>  static int
> -i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring,
> +i915_gem_execbuffer_reserve(struct intel_engine *ring,
>  			    struct list_head *vmas,
>  			    bool *need_relocs)
>  {
> @@ -711,7 +711,7 @@ static int
>  i915_gem_execbuffer_relocate_slow(struct drm_device *dev,
>  				  struct drm_i915_gem_execbuffer2 *args,
>  				  struct drm_file *file,
> -				  struct intel_ring_buffer *ring,
> +				  struct intel_engine *ring,
>  				  struct eb_vmas *eb,
>  				  struct drm_i915_gem_exec_object2 *exec)
>  {
> @@ -827,7 +827,7 @@ err:
>  }
>  
>  static int
> -i915_gem_execbuffer_move_to_gpu(struct intel_ring_buffer *ring,
> +i915_gem_execbuffer_move_to_gpu(struct intel_engine *ring,
>  				struct list_head *vmas)
>  {
>  	struct i915_vma *vma;
> @@ -912,7 +912,7 @@ validate_exec_list(struct drm_i915_gem_exec_object2 *exec,
>  
>  static struct i915_hw_context *
>  i915_gem_validate_context(struct drm_device *dev, struct drm_file *file,
> -			  struct intel_ring_buffer *ring, const u32 ctx_id)
> +			  struct intel_engine *ring, const u32 ctx_id)
>  {
>  	struct i915_hw_context *ctx = NULL;
>  	struct i915_ctx_hang_stats *hs;
> @@ -935,7 +935,7 @@ i915_gem_validate_context(struct drm_device *dev, struct drm_file *file,
>  
>  static void
>  i915_gem_execbuffer_move_to_active(struct list_head *vmas,
> -				   struct intel_ring_buffer *ring)
> +				   struct intel_engine *ring)
>  {
>  	struct i915_vma *vma;
>  
> @@ -970,7 +970,7 @@ i915_gem_execbuffer_move_to_active(struct list_head *vmas,
>  static void
>  i915_gem_execbuffer_retire_commands(struct drm_device *dev,
>  				    struct drm_file *file,
> -				    struct intel_ring_buffer *ring,
> +				    struct intel_engine *ring,
>  				    struct drm_i915_gem_object *obj)
>  {
>  	/* Unconditionally force add_request to emit a full flush. */
> @@ -982,7 +982,7 @@ i915_gem_execbuffer_retire_commands(struct drm_device *dev,
>  
>  static int
>  i915_reset_gen7_sol_offsets(struct drm_device *dev,
> -			    struct intel_ring_buffer *ring)
> +			    struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	int ret, i;
> @@ -1048,7 +1048,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data,
>  	struct eb_vmas *eb;
>  	struct drm_i915_gem_object *batch_obj;
>  	struct drm_clip_rect *cliprects = NULL;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	struct i915_hw_context *ctx;
>  	struct i915_address_space *vm;
>  	const u32 ctx_id = i915_execbuffer2_get_context_id(*args);
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
> index 1dff805..31b58ee 100644
> --- a/drivers/gpu/drm/i915/i915_gem_gtt.c
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
> @@ -207,7 +207,7 @@ static gen6_gtt_pte_t iris_pte_encode(dma_addr_t addr,
>  }
>  
>  /* Broadwell Page Directory Pointer Descriptors */
> -static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry,
> +static int gen8_write_pdp(struct intel_engine *ring, unsigned entry,
>  			   uint64_t val, bool synchronous)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> @@ -237,7 +237,7 @@ static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry,
>  }
>  
>  static int gen8_mm_switch(struct i915_hw_ppgtt *ppgtt,
> -			  struct intel_ring_buffer *ring,
> +			  struct intel_engine *ring,
>  			  bool synchronous)
>  {
>  	int i, ret;
> @@ -716,7 +716,7 @@ static uint32_t get_pd_offset(struct i915_hw_ppgtt *ppgtt)
>  }
>  
>  static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt,
> -			 struct intel_ring_buffer *ring,
> +			 struct intel_engine *ring,
>  			 bool synchronous)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
> @@ -760,7 +760,7 @@ static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt,
>  }
>  
>  static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt,
> -			  struct intel_ring_buffer *ring,
> +			  struct intel_engine *ring,
>  			  bool synchronous)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
> @@ -811,7 +811,7 @@ static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt,
>  }
>  
>  static int gen6_mm_switch(struct i915_hw_ppgtt *ppgtt,
> -			  struct intel_ring_buffer *ring,
> +			  struct intel_engine *ring,
>  			  bool synchronous)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
> @@ -832,7 +832,7 @@ static int gen8_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int j, ret;
>  
>  	for_each_active_ring(ring, dev_priv, j) {
> @@ -862,7 +862,7 @@ static int gen7_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	uint32_t ecochk, ecobits;
>  	int i;
>  
> @@ -901,7 +901,7 @@ static int gen6_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
>  {
>  	struct drm_device *dev = ppgtt->base.dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	uint32_t ecochk, gab_ctl, ecobits;
>  	int i;
>  
> @@ -1269,7 +1269,7 @@ static void undo_idling(struct drm_i915_private *dev_priv, bool interruptible)
>  void i915_check_and_clear_faults(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	if (INTEL_INFO(dev)->gen < 6)
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
> index cfca023..0775662 100644
> --- a/drivers/gpu/drm/i915/i915_gem_gtt.h
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
> @@ -261,7 +261,7 @@ struct i915_hw_ppgtt {
>  
>  	int (*enable)(struct i915_hw_ppgtt *ppgtt);
>  	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
> -			 struct intel_ring_buffer *ring,
> +			 struct intel_engine *ring,
>  			 bool synchronous);
>  	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
>  };
> diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
> index 8f37238..0853db3 100644
> --- a/drivers/gpu/drm/i915/i915_gpu_error.c
> +++ b/drivers/gpu/drm/i915/i915_gpu_error.c
> @@ -745,7 +745,7 @@ static void i915_gem_record_fences(struct drm_device *dev,
>  }
>  
>  static void i915_record_ring_state(struct drm_device *dev,
> -				   struct intel_ring_buffer *ring,
> +				   struct intel_engine *ring,
>  				   struct drm_i915_error_ring *ering)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -857,7 +857,7 @@ static void i915_record_ring_state(struct drm_device *dev,
>  }
>  
>  
> -static void i915_gem_record_active_context(struct intel_ring_buffer *ring,
> +static void i915_gem_record_active_context(struct intel_engine *ring,
>  					   struct drm_i915_error_state *error,
>  					   struct drm_i915_error_ring *ering)
>  {
> @@ -884,7 +884,7 @@ static void i915_gem_record_rings(struct drm_device *dev,
>  	int i, count;
>  
>  	for (i = 0; i < I915_NUM_RINGS; i++) {
> -		struct intel_ring_buffer *ring = &dev_priv->ring[i];
> +		struct intel_engine *ring = &dev_priv->ring[i];
>  
>  		if (!intel_ring_initialized(ring))
>  			continue;
> diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
> index 4a8e8cb..58c8812 100644
> --- a/drivers/gpu/drm/i915/i915_irq.c
> +++ b/drivers/gpu/drm/i915/i915_irq.c
> @@ -1077,7 +1077,7 @@ static void ironlake_rps_change_irq_handler(struct drm_device *dev)
>  }
>  
>  static void notify_ring(struct drm_device *dev,
> -			struct intel_ring_buffer *ring)
> +			struct intel_engine *ring)
>  {
>  	if (ring->obj == NULL)
>  		return;
> @@ -2111,7 +2111,7 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg)
>  static void i915_error_wake_up(struct drm_i915_private *dev_priv,
>  			       bool reset_completed)
>  {
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	/*
> @@ -2544,14 +2544,14 @@ static void gen8_disable_vblank(struct drm_device *dev, int pipe)
>  }
>  
>  static u32
> -ring_last_seqno(struct intel_ring_buffer *ring)
> +ring_last_seqno(struct intel_engine *ring)
>  {
>  	return list_entry(ring->request_list.prev,
>  			  struct drm_i915_gem_request, list)->seqno;
>  }
>  
>  static bool
> -ring_idle(struct intel_ring_buffer *ring, u32 seqno)
> +ring_idle(struct intel_engine *ring, u32 seqno)
>  {
>  	return (list_empty(&ring->request_list) ||
>  		i915_seqno_passed(seqno, ring_last_seqno(ring)));
> @@ -2574,11 +2574,11 @@ ipehr_is_semaphore_wait(struct drm_device *dev, u32 ipehr)
>  	}
>  }
>  
> -static struct intel_ring_buffer *
> -semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr)
> +static struct intel_engine *
> +semaphore_wait_to_signaller_ring(struct intel_engine *ring, u32 ipehr)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> -	struct intel_ring_buffer *signaller;
> +	struct intel_engine *signaller;
>  	int i;
>  
>  	if (INTEL_INFO(dev_priv->dev)->gen >= 8) {
> @@ -2606,8 +2606,8 @@ semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr)
>  	return NULL;
>  }
>  
> -static struct intel_ring_buffer *
> -semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno)
> +static struct intel_engine *
> +semaphore_waits_for(struct intel_engine *ring, u32 *seqno)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  	u32 cmd, ipehr, head;
> @@ -2649,10 +2649,10 @@ semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno)
>  	return semaphore_wait_to_signaller_ring(ring, ipehr);
>  }
>  
> -static int semaphore_passed(struct intel_ring_buffer *ring)
> +static int semaphore_passed(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> -	struct intel_ring_buffer *signaller;
> +	struct intel_engine *signaller;
>  	u32 seqno, ctl;
>  
>  	ring->hangcheck.deadlock = true;
> @@ -2671,7 +2671,7 @@ static int semaphore_passed(struct intel_ring_buffer *ring)
>  
>  static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv)
>  {
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  
>  	for_each_active_ring(ring, dev_priv, i)
> @@ -2679,7 +2679,7 @@ static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv)
>  }
>  
>  static enum intel_ring_hangcheck_action
> -ring_stuck(struct intel_ring_buffer *ring, u64 acthd)
> +ring_stuck(struct intel_engine *ring, u64 acthd)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -2735,7 +2735,7 @@ static void i915_hangcheck_elapsed(unsigned long data)
>  {
>  	struct drm_device *dev = (struct drm_device *)data;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	int i;
>  	int busy_count = 0, rings_hung = 0;
>  	bool stuck[I915_NUM_RINGS] = { 0 };
> diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
> index b29d7b1..a4f9e62 100644
> --- a/drivers/gpu/drm/i915/i915_trace.h
> +++ b/drivers/gpu/drm/i915/i915_trace.h
> @@ -326,8 +326,8 @@ TRACE_EVENT(i915_gem_evict_vm,
>  );
>  
>  TRACE_EVENT(i915_gem_ring_sync_to,
> -	    TP_PROTO(struct intel_ring_buffer *from,
> -		     struct intel_ring_buffer *to,
> +	    TP_PROTO(struct intel_engine *from,
> +		     struct intel_engine *to,
>  		     u32 seqno),
>  	    TP_ARGS(from, to, seqno),
>  
> @@ -352,7 +352,7 @@ TRACE_EVENT(i915_gem_ring_sync_to,
>  );
>  
>  TRACE_EVENT(i915_gem_ring_dispatch,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno, u32 flags),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno, u32 flags),
>  	    TP_ARGS(ring, seqno, flags),
>  
>  	    TP_STRUCT__entry(
> @@ -375,7 +375,7 @@ TRACE_EVENT(i915_gem_ring_dispatch,
>  );
>  
>  TRACE_EVENT(i915_gem_ring_flush,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 invalidate, u32 flush),
> +	    TP_PROTO(struct intel_engine *ring, u32 invalidate, u32 flush),
>  	    TP_ARGS(ring, invalidate, flush),
>  
>  	    TP_STRUCT__entry(
> @@ -398,7 +398,7 @@ TRACE_EVENT(i915_gem_ring_flush,
>  );
>  
>  DECLARE_EVENT_CLASS(i915_gem_request,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno),
>  	    TP_ARGS(ring, seqno),
>  
>  	    TP_STRUCT__entry(
> @@ -418,12 +418,12 @@ DECLARE_EVENT_CLASS(i915_gem_request,
>  );
>  
>  DEFINE_EVENT(i915_gem_request, i915_gem_request_add,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno),
>  	    TP_ARGS(ring, seqno)
>  );
>  
>  TRACE_EVENT(i915_gem_request_complete,
> -	    TP_PROTO(struct intel_ring_buffer *ring),
> +	    TP_PROTO(struct intel_engine *ring),
>  	    TP_ARGS(ring),
>  
>  	    TP_STRUCT__entry(
> @@ -443,12 +443,12 @@ TRACE_EVENT(i915_gem_request_complete,
>  );
>  
>  DEFINE_EVENT(i915_gem_request, i915_gem_request_retire,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno),
>  	    TP_ARGS(ring, seqno)
>  );
>  
>  TRACE_EVENT(i915_gem_request_wait_begin,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno),
>  	    TP_ARGS(ring, seqno),
>  
>  	    TP_STRUCT__entry(
> @@ -477,12 +477,12 @@ TRACE_EVENT(i915_gem_request_wait_begin,
>  );
>  
>  DEFINE_EVENT(i915_gem_request, i915_gem_request_wait_end,
> -	    TP_PROTO(struct intel_ring_buffer *ring, u32 seqno),
> +	    TP_PROTO(struct intel_engine *ring, u32 seqno),
>  	    TP_ARGS(ring, seqno)
>  );
>  
>  DECLARE_EVENT_CLASS(i915_ring,
> -	    TP_PROTO(struct intel_ring_buffer *ring),
> +	    TP_PROTO(struct intel_engine *ring),
>  	    TP_ARGS(ring),
>  
>  	    TP_STRUCT__entry(
> @@ -499,12 +499,12 @@ DECLARE_EVENT_CLASS(i915_ring,
>  );
>  
>  DEFINE_EVENT(i915_ring, i915_ring_wait_begin,
> -	    TP_PROTO(struct intel_ring_buffer *ring),
> +	    TP_PROTO(struct intel_engine *ring),
>  	    TP_ARGS(ring)
>  );
>  
>  DEFINE_EVENT(i915_ring, i915_ring_wait_end,
> -	    TP_PROTO(struct intel_ring_buffer *ring),
> +	    TP_PROTO(struct intel_engine *ring),
>  	    TP_ARGS(ring)
>  );
>  
> diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
> index c65e7f7..f821147 100644
> --- a/drivers/gpu/drm/i915/intel_display.c
> +++ b/drivers/gpu/drm/i915/intel_display.c
> @@ -1944,7 +1944,7 @@ static int intel_align_height(struct drm_device *dev, int height, bool tiled)
>  int
>  intel_pin_and_fence_fb_obj(struct drm_device *dev,
>  			   struct drm_i915_gem_object *obj,
> -			   struct intel_ring_buffer *pipelined)
> +			   struct intel_engine *pipelined)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	u32 alignment;
> @@ -8424,7 +8424,7 @@ out:
>  }
>  
>  void intel_mark_fb_busy(struct drm_i915_gem_object *obj,
> -			struct intel_ring_buffer *ring)
> +			struct intel_engine *ring)
>  {
>  	struct drm_device *dev = obj->base.dev;
>  	struct drm_crtc *crtc;
> @@ -8582,7 +8582,7 @@ static int intel_gen2_queue_flip(struct drm_device *dev,
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
>  	u32 flip_mask;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	ret = intel_pin_and_fence_fb_obj(dev, obj, ring);
> @@ -8627,7 +8627,7 @@ static int intel_gen3_queue_flip(struct drm_device *dev,
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
>  	u32 flip_mask;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	ret = intel_pin_and_fence_fb_obj(dev, obj, ring);
> @@ -8669,7 +8669,7 @@ static int intel_gen4_queue_flip(struct drm_device *dev,
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
>  	uint32_t pf, pipesrc;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	ret = intel_pin_and_fence_fb_obj(dev, obj, ring);
> @@ -8717,7 +8717,7 @@ static int intel_gen6_queue_flip(struct drm_device *dev,
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	uint32_t pf, pipesrc;
>  	int ret;
>  
> @@ -8762,7 +8762,7 @@ static int intel_gen7_queue_flip(struct drm_device *dev,
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
>  	struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	uint32_t plane_bit = 0;
>  	int len, ret;
>  
> diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
> index d8b540b..23b5abf 100644
> --- a/drivers/gpu/drm/i915/intel_drv.h
> +++ b/drivers/gpu/drm/i915/intel_drv.h
> @@ -694,7 +694,7 @@ int intel_pch_rawclk(struct drm_device *dev);
>  int valleyview_cur_cdclk(struct drm_i915_private *dev_priv);
>  void intel_mark_busy(struct drm_device *dev);
>  void intel_mark_fb_busy(struct drm_i915_gem_object *obj,
> -			struct intel_ring_buffer *ring);
> +			struct intel_engine *ring);
>  void intel_mark_idle(struct drm_device *dev);
>  void intel_crtc_restore_mode(struct drm_crtc *crtc);
>  void intel_crtc_update_dpms(struct drm_crtc *crtc);
> @@ -726,7 +726,7 @@ void intel_release_load_detect_pipe(struct drm_connector *connector,
>  				    struct intel_load_detect_pipe *old);
>  int intel_pin_and_fence_fb_obj(struct drm_device *dev,
>  			       struct drm_i915_gem_object *obj,
> -			       struct intel_ring_buffer *pipelined);
> +			       struct intel_engine *pipelined);
>  void intel_unpin_fb_obj(struct drm_i915_gem_object *obj);
>  struct drm_framebuffer *
>  __intel_framebuffer_create(struct drm_device *dev,
> diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c
> index d8adc91..965eec1 100644
> --- a/drivers/gpu/drm/i915/intel_overlay.c
> +++ b/drivers/gpu/drm/i915/intel_overlay.c
> @@ -213,7 +213,7 @@ static int intel_overlay_do_wait_request(struct intel_overlay *overlay,
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	BUG_ON(overlay->last_flip_req);
> @@ -236,7 +236,7 @@ static int intel_overlay_on(struct intel_overlay *overlay)
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	BUG_ON(overlay->active);
> @@ -263,7 +263,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay,
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	u32 flip_addr = overlay->flip_addr;
>  	u32 tmp;
>  	int ret;
> @@ -320,7 +320,7 @@ static int intel_overlay_off(struct intel_overlay *overlay)
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	u32 flip_addr = overlay->flip_addr;
>  	int ret;
>  
> @@ -363,7 +363,7 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay)
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	if (overlay->last_flip_req == 0)
> @@ -389,7 +389,7 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
>  {
>  	struct drm_device *dev = overlay->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	/* Only wait if there is actually an old frame to release to
> diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c
> index acfded3..17f636e 100644
> --- a/drivers/gpu/drm/i915/intel_pm.c
> +++ b/drivers/gpu/drm/i915/intel_pm.c
> @@ -3379,7 +3379,7 @@ static void parse_rp_state_cap(struct drm_i915_private *dev_priv, u32 rp_state_c
>  static void gen8_enable_rps(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	uint32_t rc6_mask = 0, rp_state_cap;
>  	int unused;
>  
> @@ -3454,7 +3454,7 @@ static void gen8_enable_rps(struct drm_device *dev)
>  static void gen6_enable_rps(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	u32 rp_state_cap;
>  	u32 gt_perf_status;
>  	u32 rc6vids, pcu_mbox = 0, rc6_mask = 0;
> @@ -3783,7 +3783,7 @@ static void valleyview_cleanup_gt_powersave(struct drm_device *dev)
>  static void valleyview_enable_rps(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	u32 gtfifodbg, val, rc6_mode = 0;
>  	int i;
>  
> @@ -3914,7 +3914,7 @@ static int ironlake_setup_rc6(struct drm_device *dev)
>  static void ironlake_enable_rc6(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	bool was_interruptible;
>  	int ret;
>  
> @@ -4426,7 +4426,7 @@ EXPORT_SYMBOL_GPL(i915_gpu_lower);
>  bool i915_gpu_busy(void)
>  {
>  	struct drm_i915_private *dev_priv;
> -	struct intel_ring_buffer *ring;
> +	struct intel_engine *ring;
>  	bool ret = false;
>  	int i;
>  
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c
> index 5d61923..4c3cc44 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.c
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.c
> @@ -40,7 +40,7 @@
>   */
>  #define CACHELINE_BYTES 64
>  
> -static inline int ring_space(struct intel_ring_buffer *ring)
> +static inline int ring_space(struct intel_engine *ring)
>  {
>  	int space = (ring->head & HEAD_ADDR) - (ring->tail + I915_RING_FREE_SPACE);
>  	if (space < 0)
> @@ -48,13 +48,13 @@ static inline int ring_space(struct intel_ring_buffer *ring)
>  	return space;
>  }
>  
> -static bool intel_ring_stopped(struct intel_ring_buffer *ring)
> +static bool intel_ring_stopped(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  	return dev_priv->gpu_error.stop_rings & intel_ring_flag(ring);
>  }
>  
> -void __intel_ring_advance(struct intel_ring_buffer *ring)
> +void __intel_ring_advance(struct intel_engine *ring)
>  {
>  	ring->tail &= ring->size - 1;
>  	if (intel_ring_stopped(ring))
> @@ -63,7 +63,7 @@ void __intel_ring_advance(struct intel_ring_buffer *ring)
>  }
>  
>  static int
> -gen2_render_ring_flush(struct intel_ring_buffer *ring,
> +gen2_render_ring_flush(struct intel_engine *ring,
>  		       u32	invalidate_domains,
>  		       u32	flush_domains)
>  {
> @@ -89,7 +89,7 @@ gen2_render_ring_flush(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -gen4_render_ring_flush(struct intel_ring_buffer *ring,
> +gen4_render_ring_flush(struct intel_engine *ring,
>  		       u32	invalidate_domains,
>  		       u32	flush_domains)
>  {
> @@ -184,7 +184,7 @@ gen4_render_ring_flush(struct intel_ring_buffer *ring,
>   * really our business.  That leaves only stall at scoreboard.
>   */
>  static int
> -intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring)
> +intel_emit_post_sync_nonzero_flush(struct intel_engine *ring)
>  {
>  	u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES;
>  	int ret;
> @@ -219,7 +219,7 @@ intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring)
>  }
>  
>  static int
> -gen6_render_ring_flush(struct intel_ring_buffer *ring,
> +gen6_render_ring_flush(struct intel_engine *ring,
>                           u32 invalidate_domains, u32 flush_domains)
>  {
>  	u32 flags = 0;
> @@ -271,7 +271,7 @@ gen6_render_ring_flush(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring)
> +gen7_render_ring_cs_stall_wa(struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -289,7 +289,7 @@ gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring)
>  	return 0;
>  }
>  
> -static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value)
> +static int gen7_ring_fbc_flush(struct intel_engine *ring, u32 value)
>  {
>  	int ret;
>  
> @@ -313,7 +313,7 @@ static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value)
>  }
>  
>  static int
> -gen7_render_ring_flush(struct intel_ring_buffer *ring,
> +gen7_render_ring_flush(struct intel_engine *ring,
>  		       u32 invalidate_domains, u32 flush_domains)
>  {
>  	u32 flags = 0;
> @@ -374,7 +374,7 @@ gen7_render_ring_flush(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -gen8_render_ring_flush(struct intel_ring_buffer *ring,
> +gen8_render_ring_flush(struct intel_engine *ring,
>  		       u32 invalidate_domains, u32 flush_domains)
>  {
>  	u32 flags = 0;
> @@ -414,14 +414,14 @@ gen8_render_ring_flush(struct intel_ring_buffer *ring,
>  
>  }
>  
> -static void ring_write_tail(struct intel_ring_buffer *ring,
> +static void ring_write_tail(struct intel_engine *ring,
>  			    u32 value)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  	I915_WRITE_TAIL(ring, value);
>  }
>  
> -u64 intel_ring_get_active_head(struct intel_ring_buffer *ring)
> +u64 intel_ring_get_active_head(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  	u64 acthd;
> @@ -437,7 +437,7 @@ u64 intel_ring_get_active_head(struct intel_ring_buffer *ring)
>  	return acthd;
>  }
>  
> -static void ring_setup_phys_status_page(struct intel_ring_buffer *ring)
> +static void ring_setup_phys_status_page(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  	u32 addr;
> @@ -448,7 +448,7 @@ static void ring_setup_phys_status_page(struct intel_ring_buffer *ring)
>  	I915_WRITE(HWS_PGA, addr);
>  }
>  
> -static bool stop_ring(struct intel_ring_buffer *ring)
> +static bool stop_ring(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = to_i915(ring->dev);
>  
> @@ -472,7 +472,7 @@ static bool stop_ring(struct intel_ring_buffer *ring)
>  	return (I915_READ_HEAD(ring) & HEAD_ADDR) == 0;
>  }
>  
> -static int init_ring_common(struct intel_ring_buffer *ring)
> +static int init_ring_common(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -550,7 +550,7 @@ out:
>  }
>  
>  static int
> -init_pipe_control(struct intel_ring_buffer *ring)
> +init_pipe_control(struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -591,7 +591,7 @@ err:
>  	return ret;
>  }
>  
> -static int init_render_ring(struct intel_ring_buffer *ring)
> +static int init_render_ring(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -647,7 +647,7 @@ static int init_render_ring(struct intel_ring_buffer *ring)
>  	return ret;
>  }
>  
> -static void render_ring_cleanup(struct intel_ring_buffer *ring)
> +static void render_ring_cleanup(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  
> @@ -663,12 +663,12 @@ static void render_ring_cleanup(struct intel_ring_buffer *ring)
>  	ring->scratch.obj = NULL;
>  }
>  
> -static int gen6_signal(struct intel_ring_buffer *signaller,
> +static int gen6_signal(struct intel_engine *signaller,
>  		       unsigned int num_dwords)
>  {
>  	struct drm_device *dev = signaller->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *useless;
> +	struct intel_engine *useless;
>  	int i, ret;
>  
>  	/* NB: In order to be able to do semaphore MBOX updates for varying
> @@ -713,7 +713,7 @@ static int gen6_signal(struct intel_ring_buffer *signaller,
>   * This acts like a signal in the canonical semaphore.
>   */
>  static int
> -gen6_add_request(struct intel_ring_buffer *ring)
> +gen6_add_request(struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -745,8 +745,8 @@ static inline bool i915_gem_has_seqno_wrapped(struct drm_device *dev,
>   * @seqno - seqno which the waiter will block on
>   */
>  static int
> -gen6_ring_sync(struct intel_ring_buffer *waiter,
> -	       struct intel_ring_buffer *signaller,
> +gen6_ring_sync(struct intel_engine *waiter,
> +	       struct intel_engine *signaller,
>  	       u32 seqno)
>  {
>  	u32 dw1 = MI_SEMAPHORE_MBOX |
> @@ -794,7 +794,7 @@ do {									\
>  } while (0)
>  
>  static int
> -pc_render_add_request(struct intel_ring_buffer *ring)
> +pc_render_add_request(struct intel_engine *ring)
>  {
>  	u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES;
>  	int ret;
> @@ -842,7 +842,7 @@ pc_render_add_request(struct intel_ring_buffer *ring)
>  }
>  
>  static u32
> -gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency)
> +gen6_ring_get_seqno(struct intel_engine *ring, bool lazy_coherency)
>  {
>  	/* Workaround to force correct ordering between irq and seqno writes on
>  	 * ivb (and maybe also on snb) by reading from a CS register (like
> @@ -856,31 +856,31 @@ gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency)
>  }
>  
>  static u32
> -ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency)
> +ring_get_seqno(struct intel_engine *ring, bool lazy_coherency)
>  {
>  	return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
>  }
>  
>  static void
> -ring_set_seqno(struct intel_ring_buffer *ring, u32 seqno)
> +ring_set_seqno(struct intel_engine *ring, u32 seqno)
>  {
>  	intel_write_status_page(ring, I915_GEM_HWS_INDEX, seqno);
>  }
>  
>  static u32
> -pc_render_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency)
> +pc_render_get_seqno(struct intel_engine *ring, bool lazy_coherency)
>  {
>  	return ring->scratch.cpu_page[0];
>  }
>  
>  static void
> -pc_render_set_seqno(struct intel_ring_buffer *ring, u32 seqno)
> +pc_render_set_seqno(struct intel_engine *ring, u32 seqno)
>  {
>  	ring->scratch.cpu_page[0] = seqno;
>  }
>  
>  static bool
> -gen5_ring_get_irq(struct intel_ring_buffer *ring)
> +gen5_ring_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -898,7 +898,7 @@ gen5_ring_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -gen5_ring_put_irq(struct intel_ring_buffer *ring)
> +gen5_ring_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -911,7 +911,7 @@ gen5_ring_put_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static bool
> -i9xx_ring_get_irq(struct intel_ring_buffer *ring)
> +i9xx_ring_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -932,7 +932,7 @@ i9xx_ring_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -i9xx_ring_put_irq(struct intel_ring_buffer *ring)
> +i9xx_ring_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -948,7 +948,7 @@ i9xx_ring_put_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static bool
> -i8xx_ring_get_irq(struct intel_ring_buffer *ring)
> +i8xx_ring_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -969,7 +969,7 @@ i8xx_ring_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -i8xx_ring_put_irq(struct intel_ring_buffer *ring)
> +i8xx_ring_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -984,7 +984,7 @@ i8xx_ring_put_irq(struct intel_ring_buffer *ring)
>  	spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
>  }
>  
> -void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
> +void intel_ring_setup_status_page(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> @@ -1047,7 +1047,7 @@ void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
>  }
>  
>  static int
> -bsd_ring_flush(struct intel_ring_buffer *ring,
> +bsd_ring_flush(struct intel_engine *ring,
>  	       u32     invalidate_domains,
>  	       u32     flush_domains)
>  {
> @@ -1064,7 +1064,7 @@ bsd_ring_flush(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -i9xx_add_request(struct intel_ring_buffer *ring)
> +i9xx_add_request(struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -1082,7 +1082,7 @@ i9xx_add_request(struct intel_ring_buffer *ring)
>  }
>  
>  static bool
> -gen6_ring_get_irq(struct intel_ring_buffer *ring)
> +gen6_ring_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1107,7 +1107,7 @@ gen6_ring_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -gen6_ring_put_irq(struct intel_ring_buffer *ring)
> +gen6_ring_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1125,7 +1125,7 @@ gen6_ring_put_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static bool
> -hsw_vebox_get_irq(struct intel_ring_buffer *ring)
> +hsw_vebox_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1145,7 +1145,7 @@ hsw_vebox_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -hsw_vebox_put_irq(struct intel_ring_buffer *ring)
> +hsw_vebox_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1163,7 +1163,7 @@ hsw_vebox_put_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static bool
> -gen8_ring_get_irq(struct intel_ring_buffer *ring)
> +gen8_ring_get_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1189,7 +1189,7 @@ gen8_ring_get_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static void
> -gen8_ring_put_irq(struct intel_ring_buffer *ring)
> +gen8_ring_put_irq(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1209,7 +1209,7 @@ gen8_ring_put_irq(struct intel_ring_buffer *ring)
>  }
>  
>  static int
> -i965_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +i965_dispatch_execbuffer(struct intel_engine *ring,
>  			 u64 offset, u32 length,
>  			 unsigned flags)
>  {
> @@ -1232,7 +1232,7 @@ i965_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  /* Just userspace ABI convention to limit the wa batch bo to a resonable size */
>  #define I830_BATCH_LIMIT (256*1024)
>  static int
> -i830_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +i830_dispatch_execbuffer(struct intel_engine *ring,
>  				u64 offset, u32 len,
>  				unsigned flags)
>  {
> @@ -1283,7 +1283,7 @@ i830_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -i915_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +i915_dispatch_execbuffer(struct intel_engine *ring,
>  			 u64 offset, u32 len,
>  			 unsigned flags)
>  {
> @@ -1300,7 +1300,7 @@ i915_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  	return 0;
>  }
>  
> -static void cleanup_status_page(struct intel_ring_buffer *ring)
> +static void cleanup_status_page(struct intel_engine *ring)
>  {
>  	struct drm_i915_gem_object *obj;
>  
> @@ -1314,7 +1314,7 @@ static void cleanup_status_page(struct intel_ring_buffer *ring)
>  	ring->status_page.obj = NULL;
>  }
>  
> -static int init_status_page(struct intel_ring_buffer *ring)
> +static int init_status_page(struct intel_engine *ring)
>  {
>  	struct drm_i915_gem_object *obj;
>  
> @@ -1351,7 +1351,7 @@ err_unref:
>  	return 0;
>  }
>  
> -static int init_phys_status_page(struct intel_ring_buffer *ring)
> +static int init_phys_status_page(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  
> @@ -1368,7 +1368,7 @@ static int init_phys_status_page(struct intel_ring_buffer *ring)
>  	return 0;
>  }
>  
> -void intel_destroy_ring_buffer(struct intel_ring_buffer *ring)
> +void intel_destroy_ring_buffer(struct intel_engine *ring)
>  {
>  	if (!ring->obj)
>  		return;
> @@ -1379,7 +1379,7 @@ void intel_destroy_ring_buffer(struct intel_ring_buffer *ring)
>  	ring->obj = NULL;
>  }
>  
> -int intel_allocate_ring_buffer(struct intel_ring_buffer *ring)
> +int intel_allocate_ring_buffer(struct intel_engine *ring)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = to_i915(dev);
> @@ -1424,7 +1424,7 @@ err_unref:
>  }
>  
>  static int intel_init_ring_buffer(struct drm_device *dev,
> -				  struct intel_ring_buffer *ring)
> +				  struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -1465,7 +1465,7 @@ static int intel_init_ring_buffer(struct drm_device *dev,
>  	return ring->init(ring);
>  }
>  
> -void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
> +void intel_cleanup_ring_buffer(struct intel_engine *ring)
>  {
>  	struct drm_i915_private *dev_priv = to_i915(ring->dev);
>  
> @@ -1485,7 +1485,7 @@ void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
>  	cleanup_status_page(ring);
>  }
>  
> -static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n)
> +static int intel_ring_wait_request(struct intel_engine *ring, int n)
>  {
>  	struct drm_i915_gem_request *request;
>  	u32 seqno = 0, tail;
> @@ -1538,7 +1538,7 @@ static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n)
>  	return 0;
>  }
>  
> -static int ring_wait_for_space(struct intel_ring_buffer *ring, int n)
> +static int ring_wait_for_space(struct intel_engine *ring, int n)
>  {
>  	struct drm_device *dev = ring->dev;
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> @@ -1586,7 +1586,7 @@ static int ring_wait_for_space(struct intel_ring_buffer *ring, int n)
>  	return -EBUSY;
>  }
>  
> -static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
> +static int intel_wrap_ring_buffer(struct intel_engine *ring)
>  {
>  	uint32_t __iomem *virt;
>  	int rem = ring->size - ring->tail;
> @@ -1608,7 +1608,7 @@ static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
>  	return 0;
>  }
>  
> -int intel_ring_idle(struct intel_ring_buffer *ring)
> +int intel_ring_idle(struct intel_engine *ring)
>  {
>  	u32 seqno;
>  	int ret;
> @@ -1632,7 +1632,7 @@ int intel_ring_idle(struct intel_ring_buffer *ring)
>  }
>  
>  static int
> -intel_ring_alloc_seqno(struct intel_ring_buffer *ring)
> +intel_ring_alloc_seqno(struct intel_engine *ring)
>  {
>  	if (ring->outstanding_lazy_seqno)
>  		return 0;
> @@ -1650,7 +1650,7 @@ intel_ring_alloc_seqno(struct intel_ring_buffer *ring)
>  	return i915_gem_get_seqno(ring->dev, &ring->outstanding_lazy_seqno);
>  }
>  
> -static int __intel_ring_prepare(struct intel_ring_buffer *ring,
> +static int __intel_ring_prepare(struct intel_engine *ring,
>  				int bytes)
>  {
>  	int ret;
> @@ -1670,7 +1670,7 @@ static int __intel_ring_prepare(struct intel_ring_buffer *ring,
>  	return 0;
>  }
>  
> -int intel_ring_begin(struct intel_ring_buffer *ring,
> +int intel_ring_begin(struct intel_engine *ring,
>  		     int num_dwords)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> @@ -1695,7 +1695,7 @@ int intel_ring_begin(struct intel_ring_buffer *ring,
>  }
>  
>  /* Align the ring tail to a cacheline boundary */
> -int intel_ring_cacheline_align(struct intel_ring_buffer *ring)
> +int intel_ring_cacheline_align(struct intel_engine *ring)
>  {
>  	int num_dwords = (ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t);
>  	int ret;
> @@ -1716,7 +1716,7 @@ int intel_ring_cacheline_align(struct intel_ring_buffer *ring)
>  	return 0;
>  }
>  
> -void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno)
> +void intel_ring_init_seqno(struct intel_engine *ring, u32 seqno)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
>  
> @@ -1733,7 +1733,7 @@ void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno)
>  	ring->hangcheck.seqno = seqno;
>  }
>  
> -static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
> +static void gen6_bsd_ring_write_tail(struct intel_engine *ring,
>  				     u32 value)
>  {
>  	struct drm_i915_private *dev_priv = ring->dev->dev_private;
> @@ -1766,7 +1766,7 @@ static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
>  		   _MASKED_BIT_DISABLE(GEN6_BSD_SLEEP_MSG_DISABLE));
>  }
>  
> -static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring,
> +static int gen6_bsd_ring_flush(struct intel_engine *ring,
>  			       u32 invalidate, u32 flush)
>  {
>  	uint32_t cmd;
> @@ -1802,7 +1802,7 @@ static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +gen8_ring_dispatch_execbuffer(struct intel_engine *ring,
>  			      u64 offset, u32 len,
>  			      unsigned flags)
>  {
> @@ -1826,7 +1826,7 @@ gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +hsw_ring_dispatch_execbuffer(struct intel_engine *ring,
>  			      u64 offset, u32 len,
>  			      unsigned flags)
>  {
> @@ -1847,7 +1847,7 @@ hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  }
>  
>  static int
> -gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
> +gen6_ring_dispatch_execbuffer(struct intel_engine *ring,
>  			      u64 offset, u32 len,
>  			      unsigned flags)
>  {
> @@ -1869,7 +1869,7 @@ gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
>  
>  /* Blitter support (SandyBridge+) */
>  
> -static int gen6_ring_flush(struct intel_ring_buffer *ring,
> +static int gen6_ring_flush(struct intel_engine *ring,
>  			   u32 invalidate, u32 flush)
>  {
>  	struct drm_device *dev = ring->dev;
> @@ -1912,7 +1912,7 @@ static int gen6_ring_flush(struct intel_ring_buffer *ring,
>  int intel_init_render_ring_buffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  
>  	if (INTEL_INFO(dev)->gen >= 6) {
>  		ring->add_request = gen6_add_request;
> @@ -2018,7 +2018,7 @@ int intel_init_render_ring_buffer(struct drm_device *dev)
>  int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
> +	struct intel_engine *ring = &dev_priv->ring[RCS];
>  	int ret;
>  
>  	if (INTEL_INFO(dev)->gen >= 6) {
> @@ -2081,7 +2081,7 @@ int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
>  int intel_init_bsd_ring_buffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[VCS];
> +	struct intel_engine *ring = &dev_priv->ring[VCS];
>  
>  	ring->write_tail = ring_write_tail;
>  	if (INTEL_INFO(dev)->gen >= 6) {
> @@ -2152,7 +2152,7 @@ int intel_init_bsd_ring_buffer(struct drm_device *dev)
>  int intel_init_bsd2_ring_buffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[VCS2];
> +	struct intel_engine *ring = &dev_priv->ring[VCS2];
>  
>  	if ((INTEL_INFO(dev)->gen != 8)) {
>  		DRM_ERROR("No dual-BSD ring on non-BDW machine\n");
> @@ -2196,7 +2196,7 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev)
>  int intel_init_blt_ring_buffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[BCS];
> +	struct intel_engine *ring = &dev_priv->ring[BCS];
>  
>  	ring->write_tail = ring_write_tail;
>  	ring->flush = gen6_ring_flush;
> @@ -2241,7 +2241,7 @@ int intel_init_blt_ring_buffer(struct drm_device *dev)
>  int intel_init_vebox_ring_buffer(struct drm_device *dev)
>  {
>  	struct drm_i915_private *dev_priv = dev->dev_private;
> -	struct intel_ring_buffer *ring = &dev_priv->ring[VECS];
> +	struct intel_engine *ring = &dev_priv->ring[VECS];
>  
>  	ring->write_tail = ring_write_tail;
>  	ring->flush = gen6_ring_flush;
> @@ -2279,7 +2279,7 @@ int intel_init_vebox_ring_buffer(struct drm_device *dev)
>  }
>  
>  int
> -intel_ring_flush_all_caches(struct intel_ring_buffer *ring)
> +intel_ring_flush_all_caches(struct intel_engine *ring)
>  {
>  	int ret;
>  
> @@ -2297,7 +2297,7 @@ intel_ring_flush_all_caches(struct intel_ring_buffer *ring)
>  }
>  
>  int
> -intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring)
> +intel_ring_invalidate_all_caches(struct intel_engine *ring)
>  {
>  	uint32_t flush_domains;
>  	int ret;
> @@ -2317,7 +2317,7 @@ intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring)
>  }
>  
>  void
> -intel_stop_ring_buffer(struct intel_ring_buffer *ring)
> +intel_stop_ring_buffer(struct intel_engine *ring)
>  {
>  	int ret;
>  
> diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
> index 680e451..50cc525 100644
> --- a/drivers/gpu/drm/i915/intel_ringbuffer.h
> +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
> @@ -54,7 +54,7 @@ struct intel_ring_hangcheck {
>  	bool deadlock;
>  };
>  
> -struct  intel_ring_buffer {
> +struct intel_engine {
>  	const char	*name;
>  	enum intel_ring_id {
>  		RCS = 0x0,
> @@ -90,33 +90,33 @@ struct  intel_ring_buffer {
>  	unsigned irq_refcount; /* protected by dev_priv->irq_lock */
>  	u32		irq_enable_mask;	/* bitmask to enable ring interrupt */
>  	u32		trace_irq_seqno;
> -	bool __must_check (*irq_get)(struct intel_ring_buffer *ring);
> -	void		(*irq_put)(struct intel_ring_buffer *ring);
> +	bool __must_check (*irq_get)(struct intel_engine *ring);
> +	void		(*irq_put)(struct intel_engine *ring);
>  
> -	int		(*init)(struct intel_ring_buffer *ring);
> +	int		(*init)(struct intel_engine *ring);
>  
> -	void		(*write_tail)(struct intel_ring_buffer *ring,
> +	void		(*write_tail)(struct intel_engine *ring,
>  				      u32 value);
> -	int __must_check (*flush)(struct intel_ring_buffer *ring,
> +	int __must_check (*flush)(struct intel_engine *ring,
>  				  u32	invalidate_domains,
>  				  u32	flush_domains);
> -	int		(*add_request)(struct intel_ring_buffer *ring);
> +	int		(*add_request)(struct intel_engine *ring);
>  	/* Some chipsets are not quite as coherent as advertised and need
>  	 * an expensive kick to force a true read of the up-to-date seqno.
>  	 * However, the up-to-date seqno is not always required and the last
>  	 * seen value is good enough. Note that the seqno will always be
>  	 * monotonic, even if not coherent.
>  	 */
> -	u32		(*get_seqno)(struct intel_ring_buffer *ring,
> +	u32		(*get_seqno)(struct intel_engine *ring,
>  				     bool lazy_coherency);
> -	void		(*set_seqno)(struct intel_ring_buffer *ring,
> +	void		(*set_seqno)(struct intel_engine *ring,
>  				     u32 seqno);
> -	int		(*dispatch_execbuffer)(struct intel_ring_buffer *ring,
> +	int		(*dispatch_execbuffer)(struct intel_engine *ring,
>  					       u64 offset, u32 length,
>  					       unsigned flags);
>  #define I915_DISPATCH_SECURE 0x1
>  #define I915_DISPATCH_PINNED 0x2
> -	void		(*cleanup)(struct intel_ring_buffer *ring);
> +	void		(*cleanup)(struct intel_engine *ring);
>  
>  	struct {
>  		u32	sync_seqno[I915_NUM_RINGS-1];
> @@ -129,10 +129,10 @@ struct  intel_ring_buffer {
>  		} mbox;
>  
>  		/* AKA wait() */
> -		int	(*sync_to)(struct intel_ring_buffer *ring,
> -				   struct intel_ring_buffer *to,
> +		int	(*sync_to)(struct intel_engine *ring,
> +				   struct intel_engine *to,
>  				   u32 seqno);
> -		int	(*signal)(struct intel_ring_buffer *signaller,
> +		int	(*signal)(struct intel_engine *signaller,
>  				  /* num_dwords needed by caller */
>  				  unsigned int num_dwords);
>  	} semaphore;
> @@ -210,20 +210,20 @@ struct  intel_ring_buffer {
>  };
>  
>  static inline bool
> -intel_ring_initialized(struct intel_ring_buffer *ring)
> +intel_ring_initialized(struct intel_engine *ring)
>  {
>  	return ring->obj != NULL;
>  }
>  
>  static inline unsigned
> -intel_ring_flag(struct intel_ring_buffer *ring)
> +intel_ring_flag(struct intel_engine *ring)
>  {
>  	return 1 << ring->id;
>  }
>  
>  static inline u32
> -intel_ring_sync_index(struct intel_ring_buffer *ring,
> -		      struct intel_ring_buffer *other)
> +intel_ring_sync_index(struct intel_engine *ring,
> +		      struct intel_engine *other)
>  {
>  	int idx;
>  
> @@ -241,7 +241,7 @@ intel_ring_sync_index(struct intel_ring_buffer *ring,
>  }
>  
>  static inline u32
> -intel_read_status_page(struct intel_ring_buffer *ring,
> +intel_read_status_page(struct intel_engine *ring,
>  		       int reg)
>  {
>  	/* Ensure that the compiler doesn't optimize away the load. */
> @@ -250,7 +250,7 @@ intel_read_status_page(struct intel_ring_buffer *ring,
>  }
>  
>  static inline void
> -intel_write_status_page(struct intel_ring_buffer *ring,
> +intel_write_status_page(struct intel_engine *ring,
>  			int reg, u32 value)
>  {
>  	ring->status_page.page_addr[reg] = value;
> @@ -275,27 +275,27 @@ intel_write_status_page(struct intel_ring_buffer *ring,
>  #define I915_GEM_HWS_SCRATCH_INDEX	0x30
>  #define I915_GEM_HWS_SCRATCH_ADDR (I915_GEM_HWS_SCRATCH_INDEX << MI_STORE_DWORD_INDEX_SHIFT)
>  
> -void intel_stop_ring_buffer(struct intel_ring_buffer *ring);
> -void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring);
> +void intel_stop_ring_buffer(struct intel_engine *ring);
> +void intel_cleanup_ring_buffer(struct intel_engine *ring);
>  
> -int __must_check intel_ring_begin(struct intel_ring_buffer *ring, int n);
> -int __must_check intel_ring_cacheline_align(struct intel_ring_buffer *ring);
> -static inline void intel_ring_emit(struct intel_ring_buffer *ring,
> +int __must_check intel_ring_begin(struct intel_engine *ring, int n);
> +int __must_check intel_ring_cacheline_align(struct intel_engine *ring);
> +static inline void intel_ring_emit(struct intel_engine *ring,
>  				   u32 data)
>  {
>  	iowrite32(data, ring->virtual_start + ring->tail);
>  	ring->tail += 4;
>  }
> -static inline void intel_ring_advance(struct intel_ring_buffer *ring)
> +static inline void intel_ring_advance(struct intel_engine *ring)
>  {
>  	ring->tail &= ring->size - 1;
>  }
> -void __intel_ring_advance(struct intel_ring_buffer *ring);
> +void __intel_ring_advance(struct intel_engine *ring);
>  
> -int __must_check intel_ring_idle(struct intel_ring_buffer *ring);
> -void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno);
> -int intel_ring_flush_all_caches(struct intel_ring_buffer *ring);
> -int intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring);
> +int __must_check intel_ring_idle(struct intel_engine *ring);
> +void intel_ring_init_seqno(struct intel_engine *ring, u32 seqno);
> +int intel_ring_flush_all_caches(struct intel_engine *ring);
> +int intel_ring_invalidate_all_caches(struct intel_engine *ring);
>  
>  void intel_init_rings_early(struct drm_device *dev);
>  int intel_init_render_ring_buffer(struct drm_device *dev);
> @@ -304,24 +304,24 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev);
>  int intel_init_blt_ring_buffer(struct drm_device *dev);
>  int intel_init_vebox_ring_buffer(struct drm_device *dev);
>  
> -u64 intel_ring_get_active_head(struct intel_ring_buffer *ring);
> -void intel_ring_setup_status_page(struct intel_ring_buffer *ring);
> +u64 intel_ring_get_active_head(struct intel_engine *ring);
> +void intel_ring_setup_status_page(struct intel_engine *ring);
>  
> -void intel_destroy_ring_buffer(struct intel_ring_buffer *ring);
> -int intel_allocate_ring_buffer(struct intel_ring_buffer *ring);
> +void intel_destroy_ring_buffer(struct intel_engine *ring);
> +int intel_allocate_ring_buffer(struct intel_engine *ring);
>  
> -static inline u32 intel_ring_get_tail(struct intel_ring_buffer *ring)
> +static inline u32 intel_ring_get_tail(struct intel_engine *ring)
>  {
>  	return ring->tail;
>  }
>  
> -static inline u32 intel_ring_get_seqno(struct intel_ring_buffer *ring)
> +static inline u32 intel_ring_get_seqno(struct intel_engine *ring)
>  {
>  	BUG_ON(ring->outstanding_lazy_seqno == 0);
>  	return ring->outstanding_lazy_seqno;
>  }
>  
> -static inline void i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno)
> +static inline void i915_trace_irq_get(struct intel_engine *ring, u32 seqno)
>  {
>  	if (ring->trace_irq_seqno == 0 && ring->irq_get(ring))
>  		ring->trace_irq_seqno = seqno;
> -- 
> 1.9.0
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx at lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/intel-gfx

-- 
Daniel Vetter
Software Engineer, Intel Corporation
+41 (0) 79 365 57 48 - http://blog.ffwll.ch



More information about the Intel-gfx mailing list