[Intel-gfx] [PATCH 5/5] drm/i915:Use the coarse mechanism based on drm fd to dispatch the BSD command on BDW GT3

Zhao Yakui yakui.zhao at intel.com
Wed Apr 9 03:59:56 CEST 2014


The BDW GT3 has two independent BSD rings, which can be used to process the
video commands. To be simpler, it is transparent to user-space driver/middleware.
Instead the kernel driver will decide which ring is to dispatch the BSD video
command.

As every BSD ring is powerful, it is enough to dispatch the BSD video command
based on the drm fd. In such case the different BSD ring is used for video playing
back and encoding. At the same time the coarse dispatch mechanism can help to avoid
the object synchronization between the BSD rings.

Signed-off-by: Zhao Yakui <yakui.zhao at intel.com>
---
 drivers/gpu/drm/i915/i915_dma.c            |   14 ++++++
 drivers/gpu/drm/i915/i915_drv.h            |    3 ++
 drivers/gpu/drm/i915/i915_gem_execbuffer.c |   73 +++++++++++++++++++++++++++-
 drivers/gpu/drm/i915/intel_ringbuffer.c    |    2 +
 drivers/gpu/drm/i915/intel_ringbuffer.h    |    2 +
 5 files changed, 93 insertions(+), 1 deletion(-)

diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c
index 0b38f88..8260463 100644
--- a/drivers/gpu/drm/i915/i915_dma.c
+++ b/drivers/gpu/drm/i915/i915_dma.c
@@ -1572,6 +1572,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags)
 	spin_lock_init(&dev_priv->backlight_lock);
 	spin_lock_init(&dev_priv->uncore.lock);
 	spin_lock_init(&dev_priv->mm.object_stat_lock);
+	spin_lock_init(&dev_priv->bsd_lock);
 	mutex_init(&dev_priv->dpio_lock);
 	mutex_init(&dev_priv->modeset_restore_lock);
 
@@ -1928,7 +1929,20 @@ void i915_driver_preclose(struct drm_device * dev, struct drm_file *file_priv)
 void i915_driver_postclose(struct drm_device *dev, struct drm_file *file)
 {
 	struct drm_i915_file_private *file_priv = file->driver_priv;
+	struct intel_ring_buffer *bsd_ring;
+	struct drm_i915_private *dev_priv = dev->dev_private;
 
+	if (file_priv && file_priv->bsd_ring) {
+		int cmd_counter;
+		bsd_ring = file_priv->bsd_ring;
+		file_priv->bsd_ring = NULL;
+		spin_lock(&dev_priv->bsd_lock);
+		cmd_counter = atomic_sub_return(1, &bsd_ring->bsd_cmd_counter);
+		if (cmd_counter < 0) {
+			atomic_set(&bsd_ring->bsd_cmd_counter, 0);
+		}
+		spin_unlock(&dev_priv->bsd_lock);
+	}
 	kfree(file_priv);
 }
 
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index d77f4e0..128639c 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -1457,6 +1457,8 @@ struct drm_i915_private {
 	struct i915_dri1_state dri1;
 	/* Old ums support infrastructure, same warning applies. */
 	struct i915_ums_state ums;
+	/* the lock for dispatch video commands on two BSD rings */
+	spinlock_t bsd_lock;
 };
 
 static inline struct drm_i915_private *to_i915(const struct drm_device *dev)
@@ -1664,6 +1666,7 @@ struct drm_i915_file_private {
 
 	struct i915_hw_context *private_default_ctx;
 	atomic_t rps_wait_boost;
+	struct  intel_ring_buffer *bsd_ring;
 };
 
 /*
diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
index 3491402..75d8cc0 100644
--- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
@@ -999,6 +999,70 @@ i915_reset_gen7_sol_offsets(struct drm_device *dev,
 	return 0;
 }
 
+/**
+ * Find one BSD ring to dispatch the corresponding BSD command.
+ * The Ring ID is returned.
+ */
+static int gen8_dispatch_bsd_ring(struct drm_device *dev,
+				  struct drm_file *file)
+{
+	struct drm_i915_private *dev_priv = dev->dev_private;
+	struct drm_i915_file_private *file_priv = file->driver_priv;
+	struct intel_ring_buffer *temp_ring, *bsd_ring;
+	int bsd_counter, temp_counter;
+
+	if (file_priv->bsd_ring) {
+		/* Check whether the load balance is required.*/
+		spin_lock(&dev_priv->bsd_lock);
+		bsd_counter = atomic_read(&(file_priv->bsd_ring->bsd_cmd_counter));
+		temp_ring = &dev_priv->ring[VCS];
+		temp_counter = atomic_read(&temp_ring->bsd_cmd_counter);
+		bsd_ring = &dev_priv->ring[VCS];
+
+		temp_ring = &dev_priv->ring[VCS2];
+		if (atomic_read(&temp_ring->bsd_cmd_counter) < temp_counter) {
+			temp_counter = atomic_read(&temp_ring->bsd_cmd_counter);
+			bsd_ring = temp_ring;
+		}
+		/*
+		 * If it is already the ring with the minimum load, it is
+		 * unnecessary to switch it.
+		 */
+		if (bsd_ring == file_priv->bsd_ring) {
+			spin_unlock(&dev_priv->bsd_lock);
+			return bsd_ring->id;
+		}
+		/*
+		 * If the load delta between current ring and target ring is
+		 * small, it is unnecessary to switch it.
+		 */
+		if ((bsd_counter - temp_counter) <= 1) {
+			spin_unlock(&dev_priv->bsd_lock);
+			return file_priv->bsd_ring->id;
+		}
+		/* balance the load between current ring and target ring */
+		atomic_dec(&file_priv->bsd_ring->bsd_cmd_counter);
+		atomic_inc(&bsd_ring->bsd_cmd_counter);
+		spin_unlock(&dev_priv->bsd_lock);
+		file_priv->bsd_ring = bsd_ring;
+		return bsd_ring->id;
+	} else {
+		spin_lock(&dev_priv->bsd_lock);
+		bsd_ring = &dev_priv->ring[VCS];
+		bsd_counter = atomic_read(&bsd_ring->bsd_cmd_counter);
+		temp_ring = &dev_priv->ring[VCS2];
+		temp_counter = atomic_read(&temp_ring->bsd_cmd_counter);
+		if (temp_counter < bsd_counter) {
+			bsd_ring = temp_ring;
+			bsd_counter = temp_counter;
+		}
+		atomic_inc(&bsd_ring->bsd_cmd_counter);
+		file_priv->bsd_ring = bsd_ring;
+		spin_unlock(&dev_priv->bsd_lock);
+		return bsd_ring->id;
+	}
+}
+
 static int
 i915_gem_do_execbuffer(struct drm_device *dev, void *data,
 		       struct drm_file *file,
@@ -1043,7 +1107,14 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data,
 
 	if ((args->flags & I915_EXEC_RING_MASK) == I915_EXEC_DEFAULT)
 		ring = &dev_priv->ring[RCS];
-	else
+	else if ((args->flags & I915_EXEC_RING_MASK) == I915_EXEC_BSD) {
+		if (HAS_BSD2(dev)) {
+			int ring_id;
+			ring_id = gen8_dispatch_bsd_ring(dev, file);
+			ring = &dev_priv->ring[ring_id];
+		} else
+			ring = &dev_priv->ring[VCS];
+	} else
 		ring = &dev_priv->ring[(args->flags & I915_EXEC_RING_MASK) - 1];
 
 	if (!intel_ring_initialized(ring)) {
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c
index 43e0227..852fc29 100644
--- a/drivers/gpu/drm/i915/intel_ringbuffer.c
+++ b/drivers/gpu/drm/i915/intel_ringbuffer.c
@@ -2123,6 +2123,7 @@ int intel_init_bsd_ring_buffer(struct drm_device *dev)
 		ring->dispatch_execbuffer = i965_dispatch_execbuffer;
 	}
 	ring->init = init_ring_common;
+	atomic_set(&ring->bsd_cmd_counter, 0);
 
 	return intel_init_ring_buffer(dev, ring);
 }
@@ -2170,6 +2171,7 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev)
 
 	ring->init = init_ring_common;
 
+	atomic_set(&ring->bsd_cmd_counter, 0);
 	return intel_init_ring_buffer(dev, ring);
 }
 
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
index 8ca4285..4f87b08 100644
--- a/drivers/gpu/drm/i915/intel_ringbuffer.h
+++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
@@ -196,6 +196,8 @@ struct  intel_ring_buffer {
 	 * to encode the command length in the header).
 	 */
 	u32 (*get_cmd_length_mask)(u32 cmd_header);
+
+	atomic_t bsd_cmd_counter;
 };
 
 static inline bool
-- 
1.7.10.1




More information about the Intel-gfx mailing list