[Intel-gfx] [PATCH 16/46] drm/i915/guc: Implement GuC parent-child context pin / unpin functions

Matthew Brost matthew.brost at intel.com
Mon Aug 9 18:58:23 UTC 2021


On Mon, Aug 09, 2021 at 05:17:34PM +0200, Daniel Vetter wrote:
> On Tue, Aug 03, 2021 at 03:29:13PM -0700, Matthew Brost wrote:
> > Implement GuC parent-child context pin / unpin functions in which in any
> > contexts in the relationship are pinned all the contexts are pinned. The
> > parent owns most of the pinning / unpinning process and the children
> > direct any pins / unpins to the parent.
> > 
> > Patch implements a number of unused functions that will be connected
> > later in the series.
> > 
> > Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> > ---
> >  drivers/gpu/drm/i915/gt/intel_context.c       | 187 ++++++++++++++++--
> >  drivers/gpu/drm/i915/gt/intel_context.h       |  43 +---
> >  drivers/gpu/drm/i915/gt/intel_context_types.h |   4 +-
> >  .../drm/i915/gt/intel_execlists_submission.c  |  25 ++-
> >  drivers/gpu/drm/i915/gt/intel_lrc.c           |  26 +--
> >  drivers/gpu/drm/i915/gt/intel_lrc.h           |   6 +-
> >  .../gpu/drm/i915/gt/intel_ring_submission.c   |   5 +-
> >  drivers/gpu/drm/i915/gt/mock_engine.c         |   4 +-
> >  .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 183 +++++++++++++++--
> >  9 files changed, 371 insertions(+), 112 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/i915/gt/intel_context.c b/drivers/gpu/drm/i915/gt/intel_context.c
> > index 8cb92b10b547..bb4c14656067 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_context.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_context.c
> > @@ -158,8 +158,8 @@ static void __ring_retire(struct intel_ring *ring)
> >  	intel_ring_unpin(ring);
> >  }
> >  
> > -static int intel_context_pre_pin(struct intel_context *ce,
> > -				 struct i915_gem_ww_ctx *ww)
> > +static int __intel_context_pre_pin(struct intel_context *ce,
> > +				   struct i915_gem_ww_ctx *ww)
> >  {
> >  	int err;
> >  
> > @@ -190,7 +190,7 @@ static int intel_context_pre_pin(struct intel_context *ce,
> >  	return err;
> >  }
> >  
> > -static void intel_context_post_unpin(struct intel_context *ce)
> > +static void __intel_context_post_unpin(struct intel_context *ce)
> >  {
> >  	if (ce->state)
> >  		__context_unpin_state(ce->state);
> > @@ -199,13 +199,85 @@ static void intel_context_post_unpin(struct intel_context *ce)
> >  	__ring_retire(ce->ring);
> >  }
> >  
> > -int __intel_context_do_pin_ww(struct intel_context *ce,
> > -			      struct i915_gem_ww_ctx *ww)
> > +static int intel_context_pre_pin(struct intel_context *ce,
> > +				 struct i915_gem_ww_ctx *ww)
> >  {
> > -	bool handoff = false;
> > -	void *vaddr;
> > +	struct intel_context *child;
> > +	int err, i = 0;
> > +
> > +	GEM_BUG_ON(intel_context_is_child(ce));
> > +
> > +	for_each_child(ce, child) {
> > +		err = __intel_context_pre_pin(child, ww);
> > +		if (unlikely(err))
> > +			goto unwind;
> > +		++i;
> > +	}
> > +
> > +	err = __intel_context_pre_pin(ce, ww);
> > +	if (unlikely(err))
> > +		goto unwind;
> > +
> > +	return 0;
> > +
> > +unwind:
> > +	for_each_child(ce, child) {
> > +		if (!i--)
> > +			break;
> > +		__intel_context_post_unpin(ce);
> > +	}
> > +
> > +	return err;
> > +}
> > +
> > +static void intel_context_post_unpin(struct intel_context *ce)
> > +{
> > +	struct intel_context *child;
> > +
> > +	GEM_BUG_ON(intel_context_is_child(ce));
> > +
> > +	for_each_child(ce, child)
> > +		__intel_context_post_unpin(child);
> > +
> > +	__intel_context_post_unpin(ce);
> > +}
> > +
> > +static int __do_ww_lock(struct intel_context *ce,
> > +			struct i915_gem_ww_ctx *ww)
> > +{
> > +	int err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww);
> > +
> > +	if (!err && ce->ring->vma->obj)
> > +		err = i915_gem_object_lock(ce->ring->vma->obj, ww);
> > +	if (!err && ce->state)
> > +		err = i915_gem_object_lock(ce->state->obj, ww);
> > +
> > +	return err;
> > +}
> > +
> > +static int do_ww_lock(struct intel_context *ce,
> > +		      struct i915_gem_ww_ctx *ww)
> > +{
> > +	struct intel_context *child;
> >  	int err = 0;
> >  
> > +	GEM_BUG_ON(intel_context_is_child(ce));
> > +
> > +	for_each_child(ce, child) {
> > +		err = __do_ww_lock(child, ww);
> > +		if (unlikely(err))
> > +			return err;
> > +	}
> > +
> > +	return __do_ww_lock(ce, ww);
> > +}
> > +
> > +static int __intel_context_do_pin_ww(struct intel_context *ce,
> > +				     struct i915_gem_ww_ctx *ww)
> > +{
> > +	bool handoff = false;
> > +	int err;
> > +
> >  	if (unlikely(!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))) {
> >  		err = intel_context_alloc_state(ce);
> >  		if (err)
> > @@ -217,14 +289,11 @@ int __intel_context_do_pin_ww(struct intel_context *ce,
> >  	 * refcount for __intel_context_active(), which prevent a lock
> >  	 * inversion of ce->pin_mutex vs dma_resv_lock().
> >  	 */
> > +	err = do_ww_lock(ce, ww);
> > +	if (err)
> > +		return err;
> >  
> > -	err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww);
> > -	if (!err && ce->ring->vma->obj)
> > -		err = i915_gem_object_lock(ce->ring->vma->obj, ww);
> > -	if (!err && ce->state)
> > -		err = i915_gem_object_lock(ce->state->obj, ww);
> > -	if (!err)
> > -		err = intel_context_pre_pin(ce, ww);
> > +	err = intel_context_pre_pin(ce, ww);
> >  	if (err)
> >  		return err;
> >  
> > @@ -232,7 +301,7 @@ int __intel_context_do_pin_ww(struct intel_context *ce,
> >  	if (err)
> >  		goto err_ctx_unpin;
> >  
> > -	err = ce->ops->pre_pin(ce, ww, &vaddr);
> > +	err = ce->ops->pre_pin(ce, ww);
> >  	if (err)
> >  		goto err_release;
> >  
> > @@ -250,7 +319,7 @@ int __intel_context_do_pin_ww(struct intel_context *ce,
> >  		if (unlikely(err))
> >  			goto err_unlock;
> >  
> > -		err = ce->ops->pin(ce, vaddr);
> > +		err = ce->ops->pin(ce);
> >  		if (err) {
> >  			intel_context_active_release(ce);
> >  			goto err_unlock;
> > @@ -290,7 +359,7 @@ int __intel_context_do_pin_ww(struct intel_context *ce,
> >  	return err;
> >  }
> >  
> > -int __intel_context_do_pin(struct intel_context *ce)
> > +static int __intel_context_do_pin(struct intel_context *ce)
> >  {
> >  	struct i915_gem_ww_ctx ww;
> >  	int err;
> > @@ -337,7 +406,7 @@ static void __intel_context_retire(struct i915_active *active)
> >  		 intel_context_get_avg_runtime_ns(ce));
> >  
> >  	set_bit(CONTEXT_VALID_BIT, &ce->flags);
> > -	intel_context_post_unpin(ce);
> > +	__intel_context_post_unpin(ce);
> >  	intel_context_put(ce);
> >  }
> >  
> > @@ -562,6 +631,88 @@ void intel_context_bind_parent_child(struct intel_context *parent,
> >  	child->parent = parent;
> >  }
> >  
> > +static inline int ____intel_context_pin(struct intel_context *ce)
> > +{
> > +	if (likely(intel_context_pin_if_active(ce)))
> > +		return 0;
> > +
> > +	return __intel_context_do_pin(ce);
> > +}
> > +
> > +static inline int __intel_context_pin_ww(struct intel_context *ce,
> > +					 struct i915_gem_ww_ctx *ww)
> > +{
> > +	if (likely(intel_context_pin_if_active(ce)))
> > +		return 0;
> > +
> > +	return __intel_context_do_pin_ww(ce, ww);
> > +}
> > +
> > +static inline void __intel_context_unpin(struct intel_context *ce)
> > +{
> > +	if (!ce->ops->sched_disable) {
> > +		__intel_context_do_unpin(ce, 1);
> > +	} else {
> > +		/*
> > +		 * Move ownership of this pin to the scheduling disable which is
> > +		 * an async operation. When that operation completes the above
> > +		 * intel_context_sched_disable_unpin is called potentially
> > +		 * unpinning the context.
> > +		 */
> > +		while (!atomic_add_unless(&ce->pin_count, -1, 1)) {
> > +			if (atomic_cmpxchg(&ce->pin_count, 1, 2) == 1) {
> 
> Uh man lockless algorithms.
> 
> Unless this comes:
> - with essentially an academic looking paper that describes the abstract
>   model of the lockless algorithm and proves it against the linux kernel
>   meory model.
> 
> - lockless stuff generally needs barriers, and those barriers must be all
>   documented. This means a) a comment next to each barrier in the code b)
>   pointing to its counterparty c) with the overall design also explained
>   in the kerneldoc for those datastructres.
> 
>   If you don't know where your barriers are, see above point about "it
>   should look more like an academic paper in the commit message"
> 
> - hard perf data about how this is absolutely required, based on a
>   real-world use-case (which then sometimes justifies a microbenchmark
>   metric for the details, but it always needs to be real-world based). And
>   also a throughrough explainer how the perf issue isn't fixable through
>   better design. If that's not doable, just protect the state machine with
>   a big dumb lock and move on.
> 
> - Also, because the current code is in such bad shape wrt lockless
>   algorithms and premature optimizations: Overall complexity should go
>   down (it's way too high right now), so pay down your new lockless trick
>   by removing one of the existing ones that we only have because we can.
> 
> Yes this is steep, but we're way out in the woods here and need to smoehow
> get back.

See below FIXME. At one point all of this was hidden in the backend but
the dma-resv patches that landed upstream completely broke the layering,
hence the need for the code here.

I guess I don't really understand what mean when you say lockless alg
needs barriers, if the atomic functions are not really atomic wouldn't
the world be broken?

Also here I don't think it is really as simple as grab big dump lock for
a variety of reasons, at least with the current dynamic pin / unpin code
in place. If we move a perma-pinned contexts this could be cleaned up
then.

Matt

> -Daniel
> 
> > +				ce->ops->sched_disable(ce);
> > +				break;
> > +			}
> > +		}
> > +	}
> > +}
> > +
> > +/*
> > + * FIXME: This is ugly, these branches are only needed for parallel contexts in
> > + * GuC submission. Basically the idea is if any of the contexts, that are
> > + * configured for parallel submission, are pinned all the contexts need to be
> > + * pinned in order to register these contexts with the GuC. We are adding the
> > + * layer here while it should probably be pushed to the backend via a vfunc. But
> > + * since we already have ce->pin + a layer atop it is confusing. Definitely
> > + * needs a bit of rework how to properly layer / structure this code path. What
> > + * is in place works but is not ideal.
> > + */
> > +int intel_context_pin(struct intel_context *ce)
> > +{
> > +	if (intel_context_is_child(ce)) {
> > +		if (!atomic_fetch_add(1, &ce->pin_count))
> > +			return ____intel_context_pin(ce->parent);
> > +		else
> > +			return 0;
> > +	} else {
> > +		return ____intel_context_pin(ce);
> > +	}
> > +}
> > +
> > +int intel_context_pin_ww(struct intel_context *ce,
> > +			 struct i915_gem_ww_ctx *ww)
> > +{
> > +	if (intel_context_is_child(ce)) {
> > +		if (!atomic_fetch_add(1, &ce->pin_count))
> > +			return __intel_context_pin_ww(ce->parent, ww);
> > +		else
> > +			return 0;
> > +	} else {
> > +		return __intel_context_pin_ww(ce, ww);
> > +	}
> > +}
> > +
> > +void intel_context_unpin(struct intel_context *ce)
> > +{
> > +	if (intel_context_is_child(ce)) {
> > +		if (atomic_fetch_add(-1, &ce->pin_count) == 1)
> > +			__intel_context_unpin(ce->parent);
> > +	} else {
> > +		__intel_context_unpin(ce);
> > +	}
> > +}
> > +
> >  #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
> >  #include "selftest_context.c"
> >  #endif
> > diff --git a/drivers/gpu/drm/i915/gt/intel_context.h b/drivers/gpu/drm/i915/gt/intel_context.h
> > index ad6ce5ac4824..c208691fc87d 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_context.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_context.h
> > @@ -110,31 +110,15 @@ static inline void intel_context_unlock_pinned(struct intel_context *ce)
> >  	mutex_unlock(&ce->pin_mutex);
> >  }
> >  
> > -int __intel_context_do_pin(struct intel_context *ce);
> > -int __intel_context_do_pin_ww(struct intel_context *ce,
> > -			      struct i915_gem_ww_ctx *ww);
> > -
> >  static inline bool intel_context_pin_if_active(struct intel_context *ce)
> >  {
> >  	return atomic_inc_not_zero(&ce->pin_count);
> >  }
> >  
> > -static inline int intel_context_pin(struct intel_context *ce)
> > -{
> > -	if (likely(intel_context_pin_if_active(ce)))
> > -		return 0;
> > -
> > -	return __intel_context_do_pin(ce);
> > -}
> > -
> > -static inline int intel_context_pin_ww(struct intel_context *ce,
> > -				       struct i915_gem_ww_ctx *ww)
> > -{
> > -	if (likely(intel_context_pin_if_active(ce)))
> > -		return 0;
> > +int intel_context_pin(struct intel_context *ce);
> >  
> > -	return __intel_context_do_pin_ww(ce, ww);
> > -}
> > +int intel_context_pin_ww(struct intel_context *ce,
> > +			 struct i915_gem_ww_ctx *ww);
> >  
> >  static inline void __intel_context_pin(struct intel_context *ce)
> >  {
> > @@ -146,28 +130,11 @@ void __intel_context_do_unpin(struct intel_context *ce, int sub);
> >  
> >  static inline void intel_context_sched_disable_unpin(struct intel_context *ce)
> >  {
> > +	GEM_BUG_ON(intel_context_is_child(ce));
> >  	__intel_context_do_unpin(ce, 2);
> >  }
> >  
> > -static inline void intel_context_unpin(struct intel_context *ce)
> > -{
> > -	if (!ce->ops->sched_disable) {
> > -		__intel_context_do_unpin(ce, 1);
> > -	} else {
> > -		/*
> > -		 * Move ownership of this pin to the scheduling disable which is
> > -		 * an async operation. When that operation completes the above
> > -		 * intel_context_sched_disable_unpin is called potentially
> > -		 * unpinning the context.
> > -		 */
> > -		while (!atomic_add_unless(&ce->pin_count, -1, 1)) {
> > -			if (atomic_cmpxchg(&ce->pin_count, 1, 2) == 1) {
> > -				ce->ops->sched_disable(ce);
> > -				break;
> > -			}
> > -		}
> > -	}
> > -}
> > +void intel_context_unpin(struct intel_context *ce);
> >  
> >  void intel_context_enter_engine(struct intel_context *ce);
> >  void intel_context_exit_engine(struct intel_context *ce);
> > diff --git a/drivers/gpu/drm/i915/gt/intel_context_types.h b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > index 66b22b370a72..eb82be15b7a2 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_context_types.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > @@ -39,8 +39,8 @@ struct intel_context_ops {
> >  
> >  	void (*ban)(struct intel_context *ce, struct i915_request *rq);
> >  
> > -	int (*pre_pin)(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void **vaddr);
> > -	int (*pin)(struct intel_context *ce, void *vaddr);
> > +	int (*pre_pin)(struct intel_context *ce, struct i915_gem_ww_ctx *ww);
> > +	int (*pin)(struct intel_context *ce);
> >  	void (*unpin)(struct intel_context *ce);
> >  	void (*post_unpin)(struct intel_context *ce);
> >  
> > diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > index baa1797af1c8..fc74ca28f245 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > @@ -2554,16 +2554,17 @@ static void execlists_submit_request(struct i915_request *request)
> >  static int
> >  __execlists_context_pre_pin(struct intel_context *ce,
> >  			    struct intel_engine_cs *engine,
> > -			    struct i915_gem_ww_ctx *ww, void **vaddr)
> > +			    struct i915_gem_ww_ctx *ww)
> >  {
> >  	int err;
> >  
> > -	err = lrc_pre_pin(ce, engine, ww, vaddr);
> > +	err = lrc_pre_pin(ce, engine, ww);
> >  	if (err)
> >  		return err;
> >  
> >  	if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags)) {
> > -		lrc_init_state(ce, engine, *vaddr);
> > +		lrc_init_state(ce, engine, ce->lrc_reg_state -
> > +			       LRC_STATE_OFFSET / sizeof(*ce->lrc_reg_state));
> >  
> >  		 __i915_gem_object_flush_map(ce->state->obj, 0, engine->context_size);
> >  	}
> > @@ -2572,15 +2573,14 @@ __execlists_context_pre_pin(struct intel_context *ce,
> >  }
> >  
> >  static int execlists_context_pre_pin(struct intel_context *ce,
> > -				     struct i915_gem_ww_ctx *ww,
> > -				     void **vaddr)
> > +				     struct i915_gem_ww_ctx *ww)
> >  {
> > -	return __execlists_context_pre_pin(ce, ce->engine, ww, vaddr);
> > +	return __execlists_context_pre_pin(ce, ce->engine, ww);
> >  }
> >  
> > -static int execlists_context_pin(struct intel_context *ce, void *vaddr)
> > +static int execlists_context_pin(struct intel_context *ce)
> >  {
> > -	return lrc_pin(ce, ce->engine, vaddr);
> > +	return lrc_pin(ce, ce->engine);
> >  }
> >  
> >  static int execlists_context_alloc(struct intel_context *ce)
> > @@ -3570,20 +3570,19 @@ static int virtual_context_alloc(struct intel_context *ce)
> >  }
> >  
> >  static int virtual_context_pre_pin(struct intel_context *ce,
> > -				   struct i915_gem_ww_ctx *ww,
> > -				   void **vaddr)
> > +				   struct i915_gem_ww_ctx *ww)
> >  {
> >  	struct virtual_engine *ve = container_of(ce, typeof(*ve), context);
> >  
> >  	 /* Note: we must use a real engine class for setting up reg state */
> > -	return __execlists_context_pre_pin(ce, ve->siblings[0], ww, vaddr);
> > +	return __execlists_context_pre_pin(ce, ve->siblings[0], ww);
> >  }
> >  
> > -static int virtual_context_pin(struct intel_context *ce, void *vaddr)
> > +static int virtual_context_pin(struct intel_context *ce)
> >  {
> >  	struct virtual_engine *ve = container_of(ce, typeof(*ve), context);
> >  
> > -	return lrc_pin(ce, ve->siblings[0], vaddr);
> > +	return lrc_pin(ce, ve->siblings[0]);
> >  }
> >  
> >  static void virtual_context_enter(struct intel_context *ce)
> > diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.c b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > index bb4af4977920..c466fc966005 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_lrc.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > @@ -947,30 +947,30 @@ void lrc_reset(struct intel_context *ce)
> >  int
> >  lrc_pre_pin(struct intel_context *ce,
> >  	    struct intel_engine_cs *engine,
> > -	    struct i915_gem_ww_ctx *ww,
> > -	    void **vaddr)
> > +	    struct i915_gem_ww_ctx *ww)
> >  {
> > +	void *vaddr;
> >  	GEM_BUG_ON(!ce->state);
> >  	GEM_BUG_ON(!i915_vma_is_pinned(ce->state));
> >  
> > -	*vaddr = i915_gem_object_pin_map(ce->state->obj,
> > -					 i915_coherent_map_type(ce->engine->i915,
> > -								ce->state->obj,
> > -								false) |
> > -					 I915_MAP_OVERRIDE);
> > +	vaddr = i915_gem_object_pin_map(ce->state->obj,
> > +					i915_coherent_map_type(ce->engine->i915,
> > +							       ce->state->obj,
> > +							       false) |
> > +					I915_MAP_OVERRIDE);
> >  
> > -	return PTR_ERR_OR_ZERO(*vaddr);
> > +	ce->lrc_reg_state = vaddr + LRC_STATE_OFFSET;
> > +
> > +	return PTR_ERR_OR_ZERO(vaddr);
> >  }
> >  
> >  int
> >  lrc_pin(struct intel_context *ce,
> > -	struct intel_engine_cs *engine,
> > -	void *vaddr)
> > +	struct intel_engine_cs *engine)
> >  {
> > -	ce->lrc_reg_state = vaddr + LRC_STATE_OFFSET;
> > -
> >  	if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags))
> > -		lrc_init_state(ce, engine, vaddr);
> > +		lrc_init_state(ce, engine,
> > +			       (void *)ce->lrc_reg_state - LRC_STATE_OFFSET);
> >  
> >  	ce->lrc.lrca = lrc_update_regs(ce, engine, ce->ring->tail);
> >  	return 0;
> > diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.h b/drivers/gpu/drm/i915/gt/intel_lrc.h
> > index 7f697845c4cf..837fcf00270d 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_lrc.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_lrc.h
> > @@ -38,12 +38,10 @@ void lrc_destroy(struct kref *kref);
> >  int
> >  lrc_pre_pin(struct intel_context *ce,
> >  	    struct intel_engine_cs *engine,
> > -	    struct i915_gem_ww_ctx *ww,
> > -	    void **vaddr);
> > +	    struct i915_gem_ww_ctx *ww);
> >  int
> >  lrc_pin(struct intel_context *ce,
> > -	struct intel_engine_cs *engine,
> > -	void *vaddr);
> > +	struct intel_engine_cs *engine);
> >  void lrc_unpin(struct intel_context *ce);
> >  void lrc_post_unpin(struct intel_context *ce);
> >  
> > diff --git a/drivers/gpu/drm/i915/gt/intel_ring_submission.c b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > index 2958e2fae380..f4f301bfb9f7 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > @@ -472,8 +472,7 @@ static int ring_context_init_default_state(struct intel_context *ce,
> >  }
> >  
> >  static int ring_context_pre_pin(struct intel_context *ce,
> > -				struct i915_gem_ww_ctx *ww,
> > -				void **unused)
> > +				struct i915_gem_ww_ctx *ww)
> >  {
> >  	struct i915_address_space *vm;
> >  	int err = 0;
> > @@ -576,7 +575,7 @@ static int ring_context_alloc(struct intel_context *ce)
> >  	return 0;
> >  }
> >  
> > -static int ring_context_pin(struct intel_context *ce, void *unused)
> > +static int ring_context_pin(struct intel_context *ce)
> >  {
> >  	return 0;
> >  }
> > diff --git a/drivers/gpu/drm/i915/gt/mock_engine.c b/drivers/gpu/drm/i915/gt/mock_engine.c
> > index 2c1af030310c..826b5d7a4573 100644
> > --- a/drivers/gpu/drm/i915/gt/mock_engine.c
> > +++ b/drivers/gpu/drm/i915/gt/mock_engine.c
> > @@ -167,12 +167,12 @@ static int mock_context_alloc(struct intel_context *ce)
> >  }
> >  
> >  static int mock_context_pre_pin(struct intel_context *ce,
> > -				struct i915_gem_ww_ctx *ww, void **unused)
> > +				struct i915_gem_ww_ctx *ww)
> >  {
> >  	return 0;
> >  }
> >  
> > -static int mock_context_pin(struct intel_context *ce, void *unused)
> > +static int mock_context_pin(struct intel_context *ce)
> >  {
> >  	return 0;
> >  }
> > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > index dec757d319a2..c5c73c42bcf7 100644
> > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > @@ -1905,6 +1905,7 @@ static int guc_lrc_desc_pin(struct intel_context *ce, bool loop)
> >  
> >  	GEM_BUG_ON(!engine->mask);
> >  	GEM_BUG_ON(context_guc_id_invalid(ce));
> > +	GEM_BUG_ON(intel_context_is_child(ce));
> >  
> >  	/*
> >  	 * Ensure LRC + CT vmas are is same region as write barrier is done
> > @@ -2008,15 +2009,13 @@ static int guc_lrc_desc_pin(struct intel_context *ce, bool loop)
> >  
> >  static int __guc_context_pre_pin(struct intel_context *ce,
> >  				 struct intel_engine_cs *engine,
> > -				 struct i915_gem_ww_ctx *ww,
> > -				 void **vaddr)
> > +				 struct i915_gem_ww_ctx *ww)
> >  {
> > -	return lrc_pre_pin(ce, engine, ww, vaddr);
> > +	return lrc_pre_pin(ce, engine, ww);
> >  }
> >  
> >  static int __guc_context_pin(struct intel_context *ce,
> > -			     struct intel_engine_cs *engine,
> > -			     void *vaddr)
> > +			     struct intel_engine_cs *engine)
> >  {
> >  	if (i915_ggtt_offset(ce->state) !=
> >  	    (ce->lrc.lrca & CTX_GTT_ADDRESS_MASK))
> > @@ -2027,20 +2026,33 @@ static int __guc_context_pin(struct intel_context *ce,
> >  	 * explaination of why.
> >  	 */
> >  
> > -	return lrc_pin(ce, engine, vaddr);
> > +	return lrc_pin(ce, engine);
> > +}
> > +
> > +static void __guc_context_unpin(struct intel_context *ce)
> > +{
> > +	lrc_unpin(ce);
> > +}
> > +
> > +static void __guc_context_post_unpin(struct intel_context *ce)
> > +{
> > +	lrc_post_unpin(ce);
> >  }
> >  
> >  static int guc_context_pre_pin(struct intel_context *ce,
> > -			       struct i915_gem_ww_ctx *ww,
> > -			       void **vaddr)
> > +			       struct i915_gem_ww_ctx *ww)
> >  {
> > -	return __guc_context_pre_pin(ce, ce->engine, ww, vaddr);
> > +	return __guc_context_pre_pin(ce, ce->engine, ww);
> >  }
> >  
> > -static int guc_context_pin(struct intel_context *ce, void *vaddr)
> > +static int guc_context_pin(struct intel_context *ce)
> >  {
> > -	int ret = __guc_context_pin(ce, ce->engine, vaddr);
> > +	int ret;
> >  
> > +	GEM_BUG_ON(intel_context_is_parent(ce) ||
> > +		   intel_context_is_child(ce));
> > +
> > +	ret = __guc_context_pin(ce, ce->engine);
> >  	if (likely(!ret && !intel_context_is_barrier(ce)))
> >  		intel_engine_pm_get(ce->engine);
> >  
> > @@ -2054,7 +2066,7 @@ static void guc_context_unpin(struct intel_context *ce)
> >  	GEM_BUG_ON(context_enabled(ce));
> >  
> >  	unpin_guc_id(guc, ce, true);
> > -	lrc_unpin(ce);
> > +	__guc_context_unpin(ce);
> >  
> >  	if (likely(!intel_context_is_barrier(ce)))
> >  		intel_engine_pm_put(ce->engine);
> > @@ -2062,7 +2074,141 @@ static void guc_context_unpin(struct intel_context *ce)
> >  
> >  static void guc_context_post_unpin(struct intel_context *ce)
> >  {
> > -	lrc_post_unpin(ce);
> > +	__guc_context_post_unpin(ce);
> > +}
> > +
> > +/* Future patches will use this function */
> > +__maybe_unused
> > +static int guc_parent_context_pre_pin(struct intel_context *ce,
> > +				      struct i915_gem_ww_ctx *ww)
> > +{
> > +	struct intel_context *child;
> > +	int err, i = 0, j = 0;
> > +
> > +	for_each_child(ce, child) {
> > +		err = i915_active_acquire(&child->active);
> > +		if (unlikely(err))
> > +			goto unwind_active;
> > +		++i;
> > +	}
> > +
> > +	for_each_child(ce, child) {
> > +		err = __guc_context_pre_pin(child, child->engine, ww);
> > +		if (unlikely(err))
> > +			goto unwind_pre_pin;
> > +		++j;
> > +	}
> > +
> > +	err = __guc_context_pre_pin(ce, ce->engine, ww);
> > +	if (unlikely(err))
> > +		goto unwind_pre_pin;
> > +
> > +	return 0;
> > +
> > +unwind_pre_pin:
> > +	for_each_child(ce, child) {
> > +		if (!j--)
> > +			break;
> > +		__guc_context_post_unpin(child);
> > +	}
> > +
> > +unwind_active:
> > +	for_each_child(ce, child) {
> > +		if (!i--)
> > +			break;
> > +		i915_active_release(&child->active);
> > +	}
> > +
> > +	return err;
> > +}
> > +
> > +/* Future patches will use this function */
> > +__maybe_unused
> > +static void guc_parent_context_post_unpin(struct intel_context *ce)
> > +{
> > +	struct intel_context *child;
> > +
> > +	for_each_child(ce, child)
> > +		__guc_context_post_unpin(child);
> > +	__guc_context_post_unpin(ce);
> > +
> > +	for_each_child(ce, child) {
> > +		intel_context_get(child);
> > +		i915_active_release(&child->active);
> > +		intel_context_put(child);
> > +	}
> > +}
> > +
> > +/* Future patches will use this function */
> > +__maybe_unused
> > +static int guc_parent_context_pin(struct intel_context *ce)
> > +{
> > +	int ret, i = 0, j = 0;
> > +	struct intel_context *child;
> > +	struct intel_engine_cs *engine;
> > +	intel_engine_mask_t tmp;
> > +
> > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > +
> > +	for_each_child(ce, child) {
> > +		ret = __guc_context_pin(child, child->engine);
> > +		if (unlikely(ret))
> > +			goto unwind_pin;
> > +		++i;
> > +	}
> > +	ret = __guc_context_pin(ce, ce->engine);
> > +	if (unlikely(ret))
> > +		goto unwind_pin;
> > +
> > +	for_each_child(ce, child)
> > +		if (test_bit(CONTEXT_LRCA_DIRTY, &child->flags)) {
> > +			set_bit(CONTEXT_LRCA_DIRTY, &ce->flags);
> > +			break;
> > +		}
> > +
> > +	for_each_engine_masked(engine, ce->engine->gt,
> > +			       ce->engine->mask, tmp)
> > +		intel_engine_pm_get(engine);
> > +	for_each_child(ce, child)
> > +		for_each_engine_masked(engine, child->engine->gt,
> > +				       child->engine->mask, tmp)
> > +			intel_engine_pm_get(engine);
> > +
> > +	return 0;
> > +
> > +unwind_pin:
> > +	for_each_child(ce, child) {
> > +		if (++j > i)
> > +			break;
> > +		__guc_context_unpin(child);
> > +	}
> > +
> > +	return ret;
> > +}
> > +
> > +/* Future patches will use this function */
> > +__maybe_unused
> > +static void guc_parent_context_unpin(struct intel_context *ce)
> > +{
> > +	struct intel_context *child;
> > +	struct intel_engine_cs *engine;
> > +	intel_engine_mask_t tmp;
> > +
> > +	GEM_BUG_ON(!intel_context_is_parent(ce));
> > +	GEM_BUG_ON(context_enabled(ce));
> > +
> > +	unpin_guc_id(ce_to_guc(ce), ce, true);
> > +	for_each_child(ce, child)
> > +		__guc_context_unpin(child);
> > +	__guc_context_unpin(ce);
> > +
> > +	for_each_engine_masked(engine, ce->engine->gt,
> > +			       ce->engine->mask, tmp)
> > +		intel_engine_pm_put(engine);
> > +	for_each_child(ce, child)
> > +		for_each_engine_masked(engine, child->engine->gt,
> > +				       child->engine->mask, tmp)
> > +			intel_engine_pm_put(engine);
> >  }
> >  
> >  static void __guc_context_sched_enable(struct intel_guc *guc,
> > @@ -2993,18 +3139,17 @@ static int guc_request_alloc(struct i915_request *rq)
> >  }
> >  
> >  static int guc_virtual_context_pre_pin(struct intel_context *ce,
> > -				       struct i915_gem_ww_ctx *ww,
> > -				       void **vaddr)
> > +				       struct i915_gem_ww_ctx *ww)
> >  {
> >  	struct intel_engine_cs *engine = guc_virtual_get_sibling(ce->engine, 0);
> >  
> > -	return __guc_context_pre_pin(ce, engine, ww, vaddr);
> > +	return __guc_context_pre_pin(ce, engine, ww);
> >  }
> >  
> > -static int guc_virtual_context_pin(struct intel_context *ce, void *vaddr)
> > +static int guc_virtual_context_pin(struct intel_context *ce)
> >  {
> >  	struct intel_engine_cs *engine = guc_virtual_get_sibling(ce->engine, 0);
> > -	int ret = __guc_context_pin(ce, engine, vaddr);
> > +	int ret = __guc_context_pin(ce, engine);
> >  	intel_engine_mask_t tmp, mask = ce->engine->mask;
> >  
> >  	if (likely(!ret))
> > @@ -3024,7 +3169,7 @@ static void guc_virtual_context_unpin(struct intel_context *ce)
> >  	GEM_BUG_ON(intel_context_is_barrier(ce));
> >  
> >  	unpin_guc_id(guc, ce, true);
> > -	lrc_unpin(ce);
> > +	__guc_context_unpin(ce);
> >  
> >  	for_each_engine_masked(engine, ce->engine->gt, mask, tmp)
> >  		intel_engine_pm_put(engine);
> > -- 
> > 2.28.0
> > 
> 
> -- 
> Daniel Vetter
> Software Engineer, Intel Corporation
> http://blog.ffwll.ch


More information about the dri-devel mailing list