[Intel-gfx] [RFC 13/15] drm/i915: Introduce intel_schedule_vblank_job()
Ville Syrjälä
ville.syrjala at linux.intel.com
Thu May 21 06:20:04 PDT 2015
On Wed, May 20, 2015 at 07:12:25PM -0700, Matt Roper wrote:
> Various places in the driver need the ability to schedule actions to run
> on a future vblank (updating watermarks, unpinning buffers, etc.). The
> long term goal is to add such a mechanism in the DRM core that can be
> shared by all drivers. Paulo Zanoni has already written some
> preliminary patches to support this, but his work will probably take
> some time to polish and get merged since it needs to untangle the DRM
> core's vblank infrastructure. This patch is partially based on Paulo's
> work, but takes a simpler approach and just adds an i915-specific
> mechanism that can be used in the interim since we have an immediate
> need for watermark updates.
>
> Inspired-by: Paulo Zanoni <przanoni at gmail.com>
> Signed-off-by: Matt Roper <matthew.d.roper at intel.com>
I like what I had better. No needless kmallocs (this IMO is a must
requirement for any solution), no needless workqueues, uses the hw
vblank counter and absolute values so isn't racy.
> ---
> drivers/gpu/drm/i915/i915_irq.c | 117 +++++++++++++++++++++++++++++++++++
> drivers/gpu/drm/i915/intel_display.c | 8 +++
> drivers/gpu/drm/i915/intel_drv.h | 27 ++++++++
> 3 files changed, 152 insertions(+)
>
> diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
> index 557af88..fd5a795 100644
> --- a/drivers/gpu/drm/i915/i915_irq.c
> +++ b/drivers/gpu/drm/i915/i915_irq.c
> @@ -1649,8 +1649,37 @@ static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir)
> }
> }
>
> +static void process_vblank_jobs(struct drm_device *dev, enum pipe pipe)
> +{
> + struct drm_i915_private *dev_priv = dev->dev_private;
> + struct drm_crtc *crtc = dev_priv->pipe_to_crtc_mapping[pipe];
> + struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> + struct i915_vblank_job *job, *tmp;
> + u32 seq;
> +
> + seq = drm_vblank_count(dev, pipe);
> + list_for_each_entry_safe(job, tmp, &intel_crtc->vblank_jobs, link) {
> + if (seq - job->seq > (1<<23))
> + continue;
> +
> + list_del(&job->link);
> + drm_vblank_put(dev, pipe);
> +
> + if (job->wq) {
> + job->fired_seq = seq;
> + queue_work(job->wq, &job->work);
> + } else {
> + job->callback(intel_crtc, job->callback_data,
> + false, seq);
> + kfree(job);
> + }
> + }
> +}
> +
> static bool intel_pipe_handle_vblank(struct drm_device *dev, enum pipe pipe)
> {
> + process_vblank_jobs(dev, pipe);
> +
> if (!drm_handle_vblank(dev, pipe))
> return false;
>
> @@ -4511,3 +4540,91 @@ void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv)
> dev_priv->dev->driver->irq_preinstall(dev_priv->dev);
> dev_priv->dev->driver->irq_postinstall(dev_priv->dev);
> }
> +
> +static void vblank_job_handler(struct work_struct *work)
> +{
> + struct i915_vblank_job *job =
> + container_of(work, struct i915_vblank_job, work);
> +
> + job->callback(job->crtc, job->callback_data,
> + job->seq != job->fired_seq, job->fired_seq);
> + kfree(job);
> +}
> +
> +/**
> + * intel_schedule_vblank_job - schedule work to happen on specified vblank
> + * @crtc: CRTC whose vblank should trigger the work
> + * @callback: Code to run on vblank
> + * @callback_data: Data to pass to callback function
> + * @wq: Workqueue to run job on (or NULL to run from interrupt context)
> + * @seq: vblank count relative to current to schedule for
> + *
> + * Schedule code that should be run after the specified vblank fires. The code
> + * can either run directly in the interrupt context or on a specified
> + * workqueue.
> + */
> +int intel_schedule_vblank_job(struct intel_crtc *crtc,
> + i915_vblank_callback_t callback,
> + void *callback_data,
> + struct workqueue_struct *wq,
> + u32 seq)
> +{
> + struct drm_device *dev = crtc->base.dev;
> + struct i915_vblank_job *job;
> + unsigned long irqflags;
> + int ret;
> +
> + job = kzalloc(sizeof(*job), GFP_ATOMIC);
> + if (!job)
> + return -ENOMEM;
> +
> + job->wq = wq;
> + job->crtc = crtc;
> + job->callback = callback;
> + job->callback_data = callback_data;
> + if (job->wq)
> + INIT_WORK(&job->work, vblank_job_handler);
> +
> + ret = drm_crtc_vblank_get(&crtc->base);
> + if (ret) {
> + DRM_DEBUG_KMS("Failed to enable interrupts, ret=%d\n", ret);
> + kfree(job);
> + return -EINVAL;
> + }
> +
> + spin_lock_irqsave(&dev->event_lock, irqflags);
> + job->seq = seq + drm_vblank_count(dev, crtc->pipe);
> + list_add_tail(&job->link, &crtc->vblank_jobs);
> + spin_unlock_irqrestore(&dev->event_lock, irqflags);
> +
> + return 0;
> +}
> +
> +/**
> + * intel_trigger_all_vblank_jobs - immediately trigger vblank jobs for a crtc
> + * @crtc: CRTC to trigger all outstanding jobs for
> + *
> + * Immediately triggers all of the jobs awaiting future vblanks on a CRTC.
> + * This will be called when a CRTC is disabled (because there may never be
> + * another vblank event). The job callbacks will receive 'true' for the
> + * early parameter, as well as the current vblank count.
> + */
> +void trigger_all_vblank_jobs(struct intel_crtc *crtc)
> +{
> + struct i915_vblank_job *job, *tmp;
> + u32 seq;
> +
> + seq = drm_vblank_count(crtc->base.dev, crtc->pipe);
> + list_for_each_entry_safe(job, tmp, &crtc->vblank_jobs, link) {
> + list_del(&job->link);
> + drm_vblank_put(crtc->base.dev, crtc->pipe);
> +
> + if (job->wq) {
> + job->fired_seq = seq;
> + queue_work(job->wq, &job->work);
> + } else {
> + job->callback(crtc, job->callback_data, true, seq);
> + kfree(job);
> + }
> + }
> +}
> diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
> index 76affba..9b56d07 100644
> --- a/drivers/gpu/drm/i915/intel_display.c
> +++ b/drivers/gpu/drm/i915/intel_display.c
> @@ -5099,6 +5099,8 @@ static void ironlake_crtc_disable(struct drm_crtc *crtc)
> if (!intel_crtc->active)
> return;
>
> + trigger_all_vblank_jobs(intel_crtc);
> +
> for_each_encoder_on_crtc(dev, crtc, encoder)
> encoder->disable(encoder);
>
> @@ -5161,6 +5163,8 @@ static void haswell_crtc_disable(struct drm_crtc *crtc)
> if (!intel_crtc->active)
> return;
>
> + trigger_all_vblank_jobs(intel_crtc);
> +
> for_each_encoder_on_crtc(dev, crtc, encoder) {
> intel_opregion_notify_encoder(encoder, false);
> encoder->disable(encoder);
> @@ -6004,6 +6008,8 @@ static void i9xx_crtc_disable(struct drm_crtc *crtc)
> if (!intel_crtc->active)
> return;
>
> + trigger_all_vblank_jobs(intel_crtc);
> +
> /*
> * On gen2 planes are double buffered but the pipe isn't, so we must
> * wait for planes to fully turn off before disabling the pipe.
> @@ -13649,6 +13655,8 @@ static void intel_crtc_init(struct drm_device *dev, int pipe)
>
> drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs);
>
> + INIT_LIST_HEAD(&intel_crtc->vblank_jobs);
> +
> WARN_ON(drm_crtc_index(&intel_crtc->base) != intel_crtc->pipe);
> return;
>
> diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
> index 627741a..630e7c1 100644
> --- a/drivers/gpu/drm/i915/intel_drv.h
> +++ b/drivers/gpu/drm/i915/intel_drv.h
> @@ -504,6 +504,24 @@ struct intel_crtc_atomic_commit {
> unsigned update_sprite_watermarks;
> };
>
> +typedef void (*i915_vblank_callback_t)(struct intel_crtc *crtc,
> + void *data,
> + bool early,
> + u32 fired_seq);
> +
> +/* Task to run (or schedule on a workqueue) on a specific vblank */
> +struct i915_vblank_job {
> + u32 seq;
> + u32 fired_seq; /* early if crtc gets disabled */
> + struct intel_crtc *crtc;
> + struct workqueue_struct *wq; /* NULL = run from interrupt */
> + i915_vblank_callback_t callback;
> + void *callback_data;
> +
> + struct list_head link;
> + struct work_struct work;
> +};
> +
> struct intel_crtc {
> struct drm_crtc base;
> enum pipe pipe;
> @@ -550,6 +568,9 @@ struct intel_crtc {
>
> /* scalers available on this crtc */
> int num_scalers;
> +
> + /* Jobs to run/schedule on vblank */
> + struct list_head vblank_jobs;
> };
>
> struct intel_plane_wm_parameters {
> @@ -913,6 +934,12 @@ static inline bool intel_irqs_enabled(struct drm_i915_private *dev_priv)
> int intel_get_crtc_scanline(struct intel_crtc *crtc);
> void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv,
> unsigned int pipe_mask);
> +int intel_schedule_vblank_job(struct intel_crtc *crtc,
> + i915_vblank_callback_t callback,
> + void *callback_data,
> + struct workqueue_struct *wq,
> + u32 seq);
> +void trigger_all_vblank_jobs(struct intel_crtc *crtc);
>
> /* intel_crt.c */
> void intel_crt_init(struct drm_device *dev);
> --
> 1.8.5.1
>
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx at lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/intel-gfx
--
Ville Syrjälä
Intel OTC
More information about the Intel-gfx
mailing list