On Thu, Aug 12, 2021 at 4:45 PM Daniel Vetter <dan...@ffwll.ch> wrote:
>
> On Wed, Aug 11, 2021 at 06:06:36PM +0000, Matthew Brost wrote:
> > On Tue, Aug 10, 2021 at 11:07:55AM +0200, Daniel Vetter wrote:
> > > On Tue, Aug 10, 2021 at 10:53:37AM +0200, Daniel Vetter wrote:
> > > > On Mon, Aug 09, 2021 at 06:58:23PM +0000, Matthew Brost wrote:
> > > > > On Mon, Aug 09, 2021 at 05:17:34PM +0200, Daniel Vetter wrote:
> > > > > > On Tue, Aug 03, 2021 at 03:29:13PM -0700, Matthew Brost wrote:
> > > > > > > Implement GuC parent-child context pin / unpin functions in which 
> > > > > > > in any
> > > > > > > contexts in the relationship are pinned all the contexts are 
> > > > > > > pinned. The
> > > > > > > parent owns most of the pinning / unpinning process and the 
> > > > > > > children
> > > > > > > direct any pins / unpins to the parent.
> > > > > > >
> > > > > > > Patch implements a number of unused functions that will be 
> > > > > > > connected
> > > > > > > later in the series.
> > > > > > >
> > > > > > > Signed-off-by: Matthew Brost <matthew.br...@intel.com>
> > > > > > > ---
> > > > > > >  drivers/gpu/drm/i915/gt/intel_context.c       | 187 
> > > > > > > ++++++++++++++++--
> > > > > > >  drivers/gpu/drm/i915/gt/intel_context.h       |  43 +---
> > > > > > >  drivers/gpu/drm/i915/gt/intel_context_types.h |   4 +-
> > > > > > >  .../drm/i915/gt/intel_execlists_submission.c  |  25 ++-
> > > > > > >  drivers/gpu/drm/i915/gt/intel_lrc.c           |  26 +--
> > > > > > >  drivers/gpu/drm/i915/gt/intel_lrc.h           |   6 +-
> > > > > > >  .../gpu/drm/i915/gt/intel_ring_submission.c   |   5 +-
> > > > > > >  drivers/gpu/drm/i915/gt/mock_engine.c         |   4 +-
> > > > > > >  .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 183 
> > > > > > > +++++++++++++++--
> > > > > > >  9 files changed, 371 insertions(+), 112 deletions(-)
> > > > > > >
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_context.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_context.c
> > > > > > > index 8cb92b10b547..bb4c14656067 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_context.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_context.c
> > > > > > > @@ -158,8 +158,8 @@ static void __ring_retire(struct intel_ring 
> > > > > > > *ring)
> > > > > > >     intel_ring_unpin(ring);
> > > > > > >  }
> > > > > > >
> > > > > > > -static int intel_context_pre_pin(struct intel_context *ce,
> > > > > > > -                            struct i915_gem_ww_ctx *ww)
> > > > > > > +static int __intel_context_pre_pin(struct intel_context *ce,
> > > > > > > +                              struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     int err;
> > > > > > >
> > > > > > > @@ -190,7 +190,7 @@ static int intel_context_pre_pin(struct 
> > > > > > > intel_context *ce,
> > > > > > >     return err;
> > > > > > >  }
> > > > > > >
> > > > > > > -static void intel_context_post_unpin(struct intel_context *ce)
> > > > > > > +static void __intel_context_post_unpin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     if (ce->state)
> > > > > > >             __context_unpin_state(ce->state);
> > > > > > > @@ -199,13 +199,85 @@ static void intel_context_post_unpin(struct 
> > > > > > > intel_context *ce)
> > > > > > >     __ring_retire(ce->ring);
> > > > > > >  }
> > > > > > >
> > > > > > > -int __intel_context_do_pin_ww(struct intel_context *ce,
> > > > > > > -                         struct i915_gem_ww_ctx *ww)
> > > > > > > +static int intel_context_pre_pin(struct intel_context *ce,
> > > > > > > +                            struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > > -   bool handoff = false;
> > > > > > > -   void *vaddr;
> > > > > > > +   struct intel_context *child;
> > > > > > > +   int err, i = 0;
> > > > > > > +
> > > > > > > +   GEM_BUG_ON(intel_context_is_child(ce));
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           err = __intel_context_pre_pin(child, ww);
> > > > > > > +           if (unlikely(err))
> > > > > > > +                   goto unwind;
> > > > > > > +           ++i;
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   err = __intel_context_pre_pin(ce, ww);
> > > > > > > +   if (unlikely(err))
> > > > > > > +           goto unwind;
> > > > > > > +
> > > > > > > +   return 0;
> > > > > > > +
> > > > > > > +unwind:
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           if (!i--)
> > > > > > > +                   break;
> > > > > > > +           __intel_context_post_unpin(ce);
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   return err;
> > > > > > > +}
> > > > > > > +
> > > > > > > +static void intel_context_post_unpin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   struct intel_context *child;
> > > > > > > +
> > > > > > > +   GEM_BUG_ON(intel_context_is_child(ce));
> > > > > > > +
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           __intel_context_post_unpin(child);
> > > > > > > +
> > > > > > > +   __intel_context_post_unpin(ce);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static int __do_ww_lock(struct intel_context *ce,
> > > > > > > +                   struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   int err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, 
> > > > > > > ww);
> > > > > > > +
> > > > > > > +   if (!err && ce->ring->vma->obj)
> > > > > > > +           err = i915_gem_object_lock(ce->ring->vma->obj, ww);
> > > > > > > +   if (!err && ce->state)
> > > > > > > +           err = i915_gem_object_lock(ce->state->obj, ww);
> > > > > > > +
> > > > > > > +   return err;
> > > > > > > +}
> > > > > > > +
> > > > > > > +static int do_ww_lock(struct intel_context *ce,
> > > > > > > +                 struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   struct intel_context *child;
> > > > > > >     int err = 0;
> > > > > > >
> > > > > > > +   GEM_BUG_ON(intel_context_is_child(ce));
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           err = __do_ww_lock(child, ww);
> > > > > > > +           if (unlikely(err))
> > > > > > > +                   return err;
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   return __do_ww_lock(ce, ww);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static int __intel_context_do_pin_ww(struct intel_context *ce,
> > > > > > > +                                struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   bool handoff = false;
> > > > > > > +   int err;
> > > > > > > +
> > > > > > >     if (unlikely(!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))) {
> > > > > > >             err = intel_context_alloc_state(ce);
> > > > > > >             if (err)
> > > > > > > @@ -217,14 +289,11 @@ int __intel_context_do_pin_ww(struct 
> > > > > > > intel_context *ce,
> > > > > > >      * refcount for __intel_context_active(), which prevent a lock
> > > > > > >      * inversion of ce->pin_mutex vs dma_resv_lock().
> > > > > > >      */
> > > > > > > +   err = do_ww_lock(ce, ww);
> > > > > > > +   if (err)
> > > > > > > +           return err;
> > > > > > >
> > > > > > > -   err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww);
> > > > > > > -   if (!err && ce->ring->vma->obj)
> > > > > > > -           err = i915_gem_object_lock(ce->ring->vma->obj, ww);
> > > > > > > -   if (!err && ce->state)
> > > > > > > -           err = i915_gem_object_lock(ce->state->obj, ww);
> > > > > > > -   if (!err)
> > > > > > > -           err = intel_context_pre_pin(ce, ww);
> > > > > > > +   err = intel_context_pre_pin(ce, ww);
> > > > > > >     if (err)
> > > > > > >             return err;
> > > > > > >
> > > > > > > @@ -232,7 +301,7 @@ int __intel_context_do_pin_ww(struct 
> > > > > > > intel_context *ce,
> > > > > > >     if (err)
> > > > > > >             goto err_ctx_unpin;
> > > > > > >
> > > > > > > -   err = ce->ops->pre_pin(ce, ww, &vaddr);
> > > > > > > +   err = ce->ops->pre_pin(ce, ww);
> > > > > > >     if (err)
> > > > > > >             goto err_release;
> > > > > > >
> > > > > > > @@ -250,7 +319,7 @@ int __intel_context_do_pin_ww(struct 
> > > > > > > intel_context *ce,
> > > > > > >             if (unlikely(err))
> > > > > > >                     goto err_unlock;
> > > > > > >
> > > > > > > -           err = ce->ops->pin(ce, vaddr);
> > > > > > > +           err = ce->ops->pin(ce);
> > > > > > >             if (err) {
> > > > > > >                     intel_context_active_release(ce);
> > > > > > >                     goto err_unlock;
> > > > > > > @@ -290,7 +359,7 @@ int __intel_context_do_pin_ww(struct 
> > > > > > > intel_context *ce,
> > > > > > >     return err;
> > > > > > >  }
> > > > > > >
> > > > > > > -int __intel_context_do_pin(struct intel_context *ce)
> > > > > > > +static int __intel_context_do_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     struct i915_gem_ww_ctx ww;
> > > > > > >     int err;
> > > > > > > @@ -337,7 +406,7 @@ static void __intel_context_retire(struct 
> > > > > > > i915_active *active)
> > > > > > >              intel_context_get_avg_runtime_ns(ce));
> > > > > > >
> > > > > > >     set_bit(CONTEXT_VALID_BIT, &ce->flags);
> > > > > > > -   intel_context_post_unpin(ce);
> > > > > > > +   __intel_context_post_unpin(ce);
> > > > > > >     intel_context_put(ce);
> > > > > > >  }
> > > > > > >
> > > > > > > @@ -562,6 +631,88 @@ void intel_context_bind_parent_child(struct 
> > > > > > > intel_context *parent,
> > > > > > >     child->parent = parent;
> > > > > > >  }
> > > > > > >
> > > > > > > +static inline int ____intel_context_pin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   if (likely(intel_context_pin_if_active(ce)))
> > > > > > > +           return 0;
> > > > > > > +
> > > > > > > +   return __intel_context_do_pin(ce);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static inline int __intel_context_pin_ww(struct intel_context 
> > > > > > > *ce,
> > > > > > > +                                    struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   if (likely(intel_context_pin_if_active(ce)))
> > > > > > > +           return 0;
> > > > > > > +
> > > > > > > +   return __intel_context_do_pin_ww(ce, ww);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static inline void __intel_context_unpin(struct intel_context 
> > > > > > > *ce)
> > > > > > > +{
> > > > > > > +   if (!ce->ops->sched_disable) {
> > > > > > > +           __intel_context_do_unpin(ce, 1);
> > > > > > > +   } else {
> > > > > > > +           /*
> > > > > > > +            * Move ownership of this pin to the scheduling 
> > > > > > > disable which is
> > > > > > > +            * an async operation. When that operation completes 
> > > > > > > the above
> > > > > > > +            * intel_context_sched_disable_unpin is called 
> > > > > > > potentially
> > > > > > > +            * unpinning the context.
> > > > > > > +            */
> > > > > > > +           while (!atomic_add_unless(&ce->pin_count, -1, 1)) {
> > > > > > > +                   if (atomic_cmpxchg(&ce->pin_count, 1, 2) == 
> > > > > > > 1) {
> > >
> > > Just as an example of what I mean here on the code review side. This is an
> > > endless loop, and you need to prove that there's no livelock or starvation
> > > issues. Or explain how else you handle that if there is one.
> > >
> >
> > If we pop into the while loop the pin_count = 1, so in all likelihood the
> > follow if evaluates to true and the loop is broken. Only way it
> > evaluates to false is the context gets pinned between the while & if, so
> > on the next pass the while statement should evaluate to false breaking
> > the looping unless of course the contexts gets unpinned again... In
> > pratice this should be at most 3 atomic operations unless the loop is
> > broken.
>
> That's not really how this works ... E.g. the linux spinlocks are
> ticketed/queued locks, exactly because the "this likely doesn't happen"
> argument is not a very good one.
>
> Can't we at least convert intel_context->pin_count into a normal counter
> with a spinlock around it and then just make all this stuff a lot more
> reasonable?
>
> Yes this doesn't help the situation overall much, but at least we're not
> spreading dubious hand-roll locking patterns all over the place for not
> much good reasons.
>
> Hand-crafted artisanal locking is never a quality sign.

Or the simple plan, just pin the contexts always. That still leaves
you with some fun on final unpin, because that's called in an awkward
place, but that can be fixed a work_struct.
-Daniel

> -Daniel
>
> >
> > Matt
> >
> > > Because unlike hand-rolled stuff linux kernel spinlocks are not dumb
> > > spinlocks, but ticketed/queued locks and therefor starvation proof. But
> > > this stuff actually matters on todays multi-core and not-so-uniform (even
> > > without fully NUMA) architectures.
> > >
> > > Also I've just found another lockless retry loop which does actually
> > > degenerate into a full endless loop (if you're sufficiently unlucky in
> > > your races), so this really isn't academic at all.
> > > -Daniel
> > >
> > > > > >
> > > > > > Uh man lockless algorithms.
> > > > > >
> > > > > > Unless this comes:
> > > > > > - with essentially an academic looking paper that describes the 
> > > > > > abstract
> > > > > >   model of the lockless algorithm and proves it against the linux 
> > > > > > kernel
> > > > > >   meory model.
> > > > > >
> > > > > > - lockless stuff generally needs barriers, and those barriers must 
> > > > > > be all
> > > > > >   documented. This means a) a comment next to each barrier in the 
> > > > > > code b)
> > > > > >   pointing to its counterparty c) with the overall design also 
> > > > > > explained
> > > > > >   in the kerneldoc for those datastructres.
> > > > > >
> > > > > >   If you don't know where your barriers are, see above point about 
> > > > > > "it
> > > > > >   should look more like an academic paper in the commit message"
> > > > > >
> > > > > > - hard perf data about how this is absolutely required, based on a
> > > > > >   real-world use-case (which then sometimes justifies a 
> > > > > > microbenchmark
> > > > > >   metric for the details, but it always needs to be real-world 
> > > > > > based). And
> > > > > >   also a throughrough explainer how the perf issue isn't fixable 
> > > > > > through
> > > > > >   better design. If that's not doable, just protect the state 
> > > > > > machine with
> > > > > >   a big dumb lock and move on.
> > > > > >
> > > > > > - Also, because the current code is in such bad shape wrt lockless
> > > > > >   algorithms and premature optimizations: Overall complexity should 
> > > > > > go
> > > > > >   down (it's way too high right now), so pay down your new lockless 
> > > > > > trick
> > > > > >   by removing one of the existing ones that we only have because we 
> > > > > > can.
> > > > > >
> > > > > > Yes this is steep, but we're way out in the woods here and need to 
> > > > > > smoehow
> > > > > > get back.
> > > > >
> > > > > See below FIXME. At one point all of this was hidden in the backend 
> > > > > but
> > > > > the dma-resv patches that landed upstream completely broke the 
> > > > > layering,
> > > > > hence the need for the code here.
> > > > >
> > > > > I guess I don't really understand what mean when you say lockless alg
> > > > > needs barriers, if the atomic functions are not really atomic wouldn't
> > > > > the world be broken?
> > > >
> > > > They unordered atomics by default. Which means they're atomic itself, 
> > > > but
> > > > entirely unordered with anything else that's going on. Except when you
> > > > have one of the atomic ops which already guarantee a barrier, or you
> > > > manually add the barriers yourself. And yes there's enormous amounts of
> > > > bugs, and with our dgpu potentially running on non-IA cpus those bugs
> > > > matter.
> > > >
> > > > Note that in C++ atomics the default behaviour is strongly ordered 
> > > > atomics
> > > > with full barriers by default, because those are much easier to program
> > > > against. Kernel isn't like that and defaults to "you need to add all the
> > > > barriers yourself".
> > > >
> > > > I have a full lenght rant in the works and will work that through all
> > > > channels, but essentially locking is really hard to get right. And
> > > > lockless tricks practically need an academic paper with a formal
> > > > correctness proof against the linux memory model, or you do have bugs.
> > > >
> > > > And I know that the current code is choke full of this stuff, so it's
> > > > tempting to just add more, but we really cant. The amount of locking
> > > > trickery we have in the codebase must go down substantially. My take is
> > > > that any code that adds anything trick needs to fully justify it against
> > > > the above list, _and_ also clean up some of the existing nonsense so 
> > > > that
> > > > overall complexity doesn't increase.
> > > >
> > > > I'll share the full length rant with you internally, it's not yet ready
> > > > for publishing (but that's planned too).
> > > >
> > > >
> > > > > Also here I don't think it is really as simple as grab big dump lock 
> > > > > for
> > > > > a variety of reasons, at least with the current dynamic pin / unpin 
> > > > > code
> > > > > in place. If we move a perma-pinned contexts this could be cleaned up
> > > > > then.
> > > >
> > > > Yes it's a disaster, but we need to stop the bleeding. If perma-pinned
> > > > context can fix this I think we should do this asap. I'd say for 
> > > > parallel
> > > > context we should just do it outright (special case them or whatever) so
> > > > that we don't have to add even more very tricky code and tech debt.
> > > >
> > > > Doable?
> > > >
> > > > Cheers, Daniel
> > > >
> > > >
> > > > >
> > > > > Matt
> > > > >
> > > > > > -Daniel
> > > > > >
> > > > > > > +                           ce->ops->sched_disable(ce);
> > > > > > > +                           break;
> > > > > > > +                   }
> > > > > > > +           }
> > > > > > > +   }
> > > > > > > +}
> > > > > > > +
> > > > > > > +/*
> > > > > > > + * FIXME: This is ugly, these branches are only needed for 
> > > > > > > parallel contexts in
> > > > > > > + * GuC submission. Basically the idea is if any of the contexts, 
> > > > > > > that are
> > > > > > > + * configured for parallel submission, are pinned all the 
> > > > > > > contexts need to be
> > > > > > > + * pinned in order to register these contexts with the GuC. We 
> > > > > > > are adding the
> > > > > > > + * layer here while it should probably be pushed to the backend 
> > > > > > > via a vfunc. But
> > > > > > > + * since we already have ce->pin + a layer atop it is confusing. 
> > > > > > > Definitely
> > > > > > > + * needs a bit of rework how to properly layer / structure this 
> > > > > > > code path. What
> > > > > > > + * is in place works but is not ideal.
> > > > > > > + */
> > > > > > > +int intel_context_pin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   if (intel_context_is_child(ce)) {
> > > > > > > +           if (!atomic_fetch_add(1, &ce->pin_count))
> > > > > > > +                   return ____intel_context_pin(ce->parent);
> > > > > > > +           else
> > > > > > > +                   return 0;
> > > > > > > +   } else {
> > > > > > > +           return ____intel_context_pin(ce);
> > > > > > > +   }
> > > > > > > +}
> > > > > > > +
> > > > > > > +int intel_context_pin_ww(struct intel_context *ce,
> > > > > > > +                    struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   if (intel_context_is_child(ce)) {
> > > > > > > +           if (!atomic_fetch_add(1, &ce->pin_count))
> > > > > > > +                   return __intel_context_pin_ww(ce->parent, ww);
> > > > > > > +           else
> > > > > > > +                   return 0;
> > > > > > > +   } else {
> > > > > > > +           return __intel_context_pin_ww(ce, ww);
> > > > > > > +   }
> > > > > > > +}
> > > > > > > +
> > > > > > > +void intel_context_unpin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   if (intel_context_is_child(ce)) {
> > > > > > > +           if (atomic_fetch_add(-1, &ce->pin_count) == 1)
> > > > > > > +                   __intel_context_unpin(ce->parent);
> > > > > > > +   } else {
> > > > > > > +           __intel_context_unpin(ce);
> > > > > > > +   }
> > > > > > > +}
> > > > > > > +
> > > > > > >  #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
> > > > > > >  #include "selftest_context.c"
> > > > > > >  #endif
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_context.h 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_context.h
> > > > > > > index ad6ce5ac4824..c208691fc87d 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_context.h
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_context.h
> > > > > > > @@ -110,31 +110,15 @@ static inline void 
> > > > > > > intel_context_unlock_pinned(struct intel_context *ce)
> > > > > > >     mutex_unlock(&ce->pin_mutex);
> > > > > > >  }
> > > > > > >
> > > > > > > -int __intel_context_do_pin(struct intel_context *ce);
> > > > > > > -int __intel_context_do_pin_ww(struct intel_context *ce,
> > > > > > > -                         struct i915_gem_ww_ctx *ww);
> > > > > > > -
> > > > > > >  static inline bool intel_context_pin_if_active(struct 
> > > > > > > intel_context *ce)
> > > > > > >  {
> > > > > > >     return atomic_inc_not_zero(&ce->pin_count);
> > > > > > >  }
> > > > > > >
> > > > > > > -static inline int intel_context_pin(struct intel_context *ce)
> > > > > > > -{
> > > > > > > -   if (likely(intel_context_pin_if_active(ce)))
> > > > > > > -           return 0;
> > > > > > > -
> > > > > > > -   return __intel_context_do_pin(ce);
> > > > > > > -}
> > > > > > > -
> > > > > > > -static inline int intel_context_pin_ww(struct intel_context *ce,
> > > > > > > -                                  struct i915_gem_ww_ctx *ww)
> > > > > > > -{
> > > > > > > -   if (likely(intel_context_pin_if_active(ce)))
> > > > > > > -           return 0;
> > > > > > > +int intel_context_pin(struct intel_context *ce);
> > > > > > >
> > > > > > > -   return __intel_context_do_pin_ww(ce, ww);
> > > > > > > -}
> > > > > > > +int intel_context_pin_ww(struct intel_context *ce,
> > > > > > > +                    struct i915_gem_ww_ctx *ww);
> > > > > > >
> > > > > > >  static inline void __intel_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > > @@ -146,28 +130,11 @@ void __intel_context_do_unpin(struct 
> > > > > > > intel_context *ce, int sub);
> > > > > > >
> > > > > > >  static inline void intel_context_sched_disable_unpin(struct 
> > > > > > > intel_context *ce)
> > > > > > >  {
> > > > > > > +   GEM_BUG_ON(intel_context_is_child(ce));
> > > > > > >     __intel_context_do_unpin(ce, 2);
> > > > > > >  }
> > > > > > >
> > > > > > > -static inline void intel_context_unpin(struct intel_context *ce)
> > > > > > > -{
> > > > > > > -   if (!ce->ops->sched_disable) {
> > > > > > > -           __intel_context_do_unpin(ce, 1);
> > > > > > > -   } else {
> > > > > > > -           /*
> > > > > > > -            * Move ownership of this pin to the scheduling 
> > > > > > > disable which is
> > > > > > > -            * an async operation. When that operation completes 
> > > > > > > the above
> > > > > > > -            * intel_context_sched_disable_unpin is called 
> > > > > > > potentially
> > > > > > > -            * unpinning the context.
> > > > > > > -            */
> > > > > > > -           while (!atomic_add_unless(&ce->pin_count, -1, 1)) {
> > > > > > > -                   if (atomic_cmpxchg(&ce->pin_count, 1, 2) == 
> > > > > > > 1) {
> > > > > > > -                           ce->ops->sched_disable(ce);
> > > > > > > -                           break;
> > > > > > > -                   }
> > > > > > > -           }
> > > > > > > -   }
> > > > > > > -}
> > > > > > > +void intel_context_unpin(struct intel_context *ce);
> > > > > > >
> > > > > > >  void intel_context_enter_engine(struct intel_context *ce);
> > > > > > >  void intel_context_exit_engine(struct intel_context *ce);
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_context_types.h 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > > > > index 66b22b370a72..eb82be15b7a2 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_context_types.h
> > > > > > > @@ -39,8 +39,8 @@ struct intel_context_ops {
> > > > > > >
> > > > > > >     void (*ban)(struct intel_context *ce, struct i915_request 
> > > > > > > *rq);
> > > > > > >
> > > > > > > -   int (*pre_pin)(struct intel_context *ce, struct 
> > > > > > > i915_gem_ww_ctx *ww, void **vaddr);
> > > > > > > -   int (*pin)(struct intel_context *ce, void *vaddr);
> > > > > > > +   int (*pre_pin)(struct intel_context *ce, struct 
> > > > > > > i915_gem_ww_ctx *ww);
> > > > > > > +   int (*pin)(struct intel_context *ce);
> > > > > > >     void (*unpin)(struct intel_context *ce);
> > > > > > >     void (*post_unpin)(struct intel_context *ce);
> > > > > > >
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > > > > > > index baa1797af1c8..fc74ca28f245 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> > > > > > > @@ -2554,16 +2554,17 @@ static void 
> > > > > > > execlists_submit_request(struct i915_request *request)
> > > > > > >  static int
> > > > > > >  __execlists_context_pre_pin(struct intel_context *ce,
> > > > > > >                         struct intel_engine_cs *engine,
> > > > > > > -                       struct i915_gem_ww_ctx *ww, void **vaddr)
> > > > > > > +                       struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     int err;
> > > > > > >
> > > > > > > -   err = lrc_pre_pin(ce, engine, ww, vaddr);
> > > > > > > +   err = lrc_pre_pin(ce, engine, ww);
> > > > > > >     if (err)
> > > > > > >             return err;
> > > > > > >
> > > > > > >     if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags)) {
> > > > > > > -           lrc_init_state(ce, engine, *vaddr);
> > > > > > > +           lrc_init_state(ce, engine, ce->lrc_reg_state -
> > > > > > > +                          LRC_STATE_OFFSET / 
> > > > > > > sizeof(*ce->lrc_reg_state));
> > > > > > >
> > > > > > >              __i915_gem_object_flush_map(ce->state->obj, 0, 
> > > > > > > engine->context_size);
> > > > > > >     }
> > > > > > > @@ -2572,15 +2573,14 @@ __execlists_context_pre_pin(struct 
> > > > > > > intel_context *ce,
> > > > > > >  }
> > > > > > >
> > > > > > >  static int execlists_context_pre_pin(struct intel_context *ce,
> > > > > > > -                                struct i915_gem_ww_ctx *ww,
> > > > > > > -                                void **vaddr)
> > > > > > > +                                struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > > -   return __execlists_context_pre_pin(ce, ce->engine, ww, vaddr);
> > > > > > > +   return __execlists_context_pre_pin(ce, ce->engine, ww);
> > > > > > >  }
> > > > > > >
> > > > > > > -static int execlists_context_pin(struct intel_context *ce, void 
> > > > > > > *vaddr)
> > > > > > > +static int execlists_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > > -   return lrc_pin(ce, ce->engine, vaddr);
> > > > > > > +   return lrc_pin(ce, ce->engine);
> > > > > > >  }
> > > > > > >
> > > > > > >  static int execlists_context_alloc(struct intel_context *ce)
> > > > > > > @@ -3570,20 +3570,19 @@ static int virtual_context_alloc(struct 
> > > > > > > intel_context *ce)
> > > > > > >  }
> > > > > > >
> > > > > > >  static int virtual_context_pre_pin(struct intel_context *ce,
> > > > > > > -                              struct i915_gem_ww_ctx *ww,
> > > > > > > -                              void **vaddr)
> > > > > > > +                              struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     struct virtual_engine *ve = container_of(ce, typeof(*ve), 
> > > > > > > context);
> > > > > > >
> > > > > > >      /* Note: we must use a real engine class for setting up reg 
> > > > > > > state */
> > > > > > > -   return __execlists_context_pre_pin(ce, ve->siblings[0], ww, 
> > > > > > > vaddr);
> > > > > > > +   return __execlists_context_pre_pin(ce, ve->siblings[0], ww);
> > > > > > >  }
> > > > > > >
> > > > > > > -static int virtual_context_pin(struct intel_context *ce, void 
> > > > > > > *vaddr)
> > > > > > > +static int virtual_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     struct virtual_engine *ve = container_of(ce, typeof(*ve), 
> > > > > > > context);
> > > > > > >
> > > > > > > -   return lrc_pin(ce, ve->siblings[0], vaddr);
> > > > > > > +   return lrc_pin(ce, ve->siblings[0]);
> > > > > > >  }
> > > > > > >
> > > > > > >  static void virtual_context_enter(struct intel_context *ce)
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > > > > > > index bb4af4977920..c466fc966005 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_lrc.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_lrc.c
> > > > > > > @@ -947,30 +947,30 @@ void lrc_reset(struct intel_context *ce)
> > > > > > >  int
> > > > > > >  lrc_pre_pin(struct intel_context *ce,
> > > > > > >         struct intel_engine_cs *engine,
> > > > > > > -       struct i915_gem_ww_ctx *ww,
> > > > > > > -       void **vaddr)
> > > > > > > +       struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > > +   void *vaddr;
> > > > > > >     GEM_BUG_ON(!ce->state);
> > > > > > >     GEM_BUG_ON(!i915_vma_is_pinned(ce->state));
> > > > > > >
> > > > > > > -   *vaddr = i915_gem_object_pin_map(ce->state->obj,
> > > > > > > -                                    
> > > > > > > i915_coherent_map_type(ce->engine->i915,
> > > > > > > -                                                           
> > > > > > > ce->state->obj,
> > > > > > > -                                                           
> > > > > > > false) |
> > > > > > > -                                    I915_MAP_OVERRIDE);
> > > > > > > +   vaddr = i915_gem_object_pin_map(ce->state->obj,
> > > > > > > +                                   
> > > > > > > i915_coherent_map_type(ce->engine->i915,
> > > > > > > +                                                          
> > > > > > > ce->state->obj,
> > > > > > > +                                                          false) 
> > > > > > > |
> > > > > > > +                                   I915_MAP_OVERRIDE);
> > > > > > >
> > > > > > > -   return PTR_ERR_OR_ZERO(*vaddr);
> > > > > > > +   ce->lrc_reg_state = vaddr + LRC_STATE_OFFSET;
> > > > > > > +
> > > > > > > +   return PTR_ERR_OR_ZERO(vaddr);
> > > > > > >  }
> > > > > > >
> > > > > > >  int
> > > > > > >  lrc_pin(struct intel_context *ce,
> > > > > > > -   struct intel_engine_cs *engine,
> > > > > > > -   void *vaddr)
> > > > > > > +   struct intel_engine_cs *engine)
> > > > > > >  {
> > > > > > > -   ce->lrc_reg_state = vaddr + LRC_STATE_OFFSET;
> > > > > > > -
> > > > > > >     if (!__test_and_set_bit(CONTEXT_INIT_BIT, &ce->flags))
> > > > > > > -           lrc_init_state(ce, engine, vaddr);
> > > > > > > +           lrc_init_state(ce, engine,
> > > > > > > +                          (void *)ce->lrc_reg_state - 
> > > > > > > LRC_STATE_OFFSET);
> > > > > > >
> > > > > > >     ce->lrc.lrca = lrc_update_regs(ce, engine, ce->ring->tail);
> > > > > > >     return 0;
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_lrc.h 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_lrc.h
> > > > > > > index 7f697845c4cf..837fcf00270d 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_lrc.h
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_lrc.h
> > > > > > > @@ -38,12 +38,10 @@ void lrc_destroy(struct kref *kref);
> > > > > > >  int
> > > > > > >  lrc_pre_pin(struct intel_context *ce,
> > > > > > >         struct intel_engine_cs *engine,
> > > > > > > -       struct i915_gem_ww_ctx *ww,
> > > > > > > -       void **vaddr);
> > > > > > > +       struct i915_gem_ww_ctx *ww);
> > > > > > >  int
> > > > > > >  lrc_pin(struct intel_context *ce,
> > > > > > > -   struct intel_engine_cs *engine,
> > > > > > > -   void *vaddr);
> > > > > > > +   struct intel_engine_cs *engine);
> > > > > > >  void lrc_unpin(struct intel_context *ce);
> > > > > > >  void lrc_post_unpin(struct intel_context *ce);
> > > > > > >
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_ring_submission.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > > > > > > index 2958e2fae380..f4f301bfb9f7 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/intel_ring_submission.c
> > > > > > > @@ -472,8 +472,7 @@ static int 
> > > > > > > ring_context_init_default_state(struct intel_context *ce,
> > > > > > >  }
> > > > > > >
> > > > > > >  static int ring_context_pre_pin(struct intel_context *ce,
> > > > > > > -                           struct i915_gem_ww_ctx *ww,
> > > > > > > -                           void **unused)
> > > > > > > +                           struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     struct i915_address_space *vm;
> > > > > > >     int err = 0;
> > > > > > > @@ -576,7 +575,7 @@ static int ring_context_alloc(struct 
> > > > > > > intel_context *ce)
> > > > > > >     return 0;
> > > > > > >  }
> > > > > > >
> > > > > > > -static int ring_context_pin(struct intel_context *ce, void 
> > > > > > > *unused)
> > > > > > > +static int ring_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     return 0;
> > > > > > >  }
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/mock_engine.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/mock_engine.c
> > > > > > > index 2c1af030310c..826b5d7a4573 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/mock_engine.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/mock_engine.c
> > > > > > > @@ -167,12 +167,12 @@ static int mock_context_alloc(struct 
> > > > > > > intel_context *ce)
> > > > > > >  }
> > > > > > >
> > > > > > >  static int mock_context_pre_pin(struct intel_context *ce,
> > > > > > > -                           struct i915_gem_ww_ctx *ww, void 
> > > > > > > **unused)
> > > > > > > +                           struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     return 0;
> > > > > > >  }
> > > > > > >
> > > > > > > -static int mock_context_pin(struct intel_context *ce, void 
> > > > > > > *unused)
> > > > > > > +static int mock_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     return 0;
> > > > > > >  }
> > > > > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 
> > > > > > > b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > > > > index dec757d319a2..c5c73c42bcf7 100644
> > > > > > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > > > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > > > > @@ -1905,6 +1905,7 @@ static int guc_lrc_desc_pin(struct 
> > > > > > > intel_context *ce, bool loop)
> > > > > > >
> > > > > > >     GEM_BUG_ON(!engine->mask);
> > > > > > >     GEM_BUG_ON(context_guc_id_invalid(ce));
> > > > > > > +   GEM_BUG_ON(intel_context_is_child(ce));
> > > > > > >
> > > > > > >     /*
> > > > > > >      * Ensure LRC + CT vmas are is same region as write barrier 
> > > > > > > is done
> > > > > > > @@ -2008,15 +2009,13 @@ static int guc_lrc_desc_pin(struct 
> > > > > > > intel_context *ce, bool loop)
> > > > > > >
> > > > > > >  static int __guc_context_pre_pin(struct intel_context *ce,
> > > > > > >                              struct intel_engine_cs *engine,
> > > > > > > -                            struct i915_gem_ww_ctx *ww,
> > > > > > > -                            void **vaddr)
> > > > > > > +                            struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > > -   return lrc_pre_pin(ce, engine, ww, vaddr);
> > > > > > > +   return lrc_pre_pin(ce, engine, ww);
> > > > > > >  }
> > > > > > >
> > > > > > >  static int __guc_context_pin(struct intel_context *ce,
> > > > > > > -                        struct intel_engine_cs *engine,
> > > > > > > -                        void *vaddr)
> > > > > > > +                        struct intel_engine_cs *engine)
> > > > > > >  {
> > > > > > >     if (i915_ggtt_offset(ce->state) !=
> > > > > > >         (ce->lrc.lrca & CTX_GTT_ADDRESS_MASK))
> > > > > > > @@ -2027,20 +2026,33 @@ static int __guc_context_pin(struct 
> > > > > > > intel_context *ce,
> > > > > > >      * explaination of why.
> > > > > > >      */
> > > > > > >
> > > > > > > -   return lrc_pin(ce, engine, vaddr);
> > > > > > > +   return lrc_pin(ce, engine);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static void __guc_context_unpin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   lrc_unpin(ce);
> > > > > > > +}
> > > > > > > +
> > > > > > > +static void __guc_context_post_unpin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   lrc_post_unpin(ce);
> > > > > > >  }
> > > > > > >
> > > > > > >  static int guc_context_pre_pin(struct intel_context *ce,
> > > > > > > -                          struct i915_gem_ww_ctx *ww,
> > > > > > > -                          void **vaddr)
> > > > > > > +                          struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > > -   return __guc_context_pre_pin(ce, ce->engine, ww, vaddr);
> > > > > > > +   return __guc_context_pre_pin(ce, ce->engine, ww);
> > > > > > >  }
> > > > > > >
> > > > > > > -static int guc_context_pin(struct intel_context *ce, void *vaddr)
> > > > > > > +static int guc_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > > -   int ret = __guc_context_pin(ce, ce->engine, vaddr);
> > > > > > > +   int ret;
> > > > > > >
> > > > > > > +   GEM_BUG_ON(intel_context_is_parent(ce) ||
> > > > > > > +              intel_context_is_child(ce));
> > > > > > > +
> > > > > > > +   ret = __guc_context_pin(ce, ce->engine);
> > > > > > >     if (likely(!ret && !intel_context_is_barrier(ce)))
> > > > > > >             intel_engine_pm_get(ce->engine);
> > > > > > >
> > > > > > > @@ -2054,7 +2066,7 @@ static void guc_context_unpin(struct 
> > > > > > > intel_context *ce)
> > > > > > >     GEM_BUG_ON(context_enabled(ce));
> > > > > > >
> > > > > > >     unpin_guc_id(guc, ce, true);
> > > > > > > -   lrc_unpin(ce);
> > > > > > > +   __guc_context_unpin(ce);
> > > > > > >
> > > > > > >     if (likely(!intel_context_is_barrier(ce)))
> > > > > > >             intel_engine_pm_put(ce->engine);
> > > > > > > @@ -2062,7 +2074,141 @@ static void guc_context_unpin(struct 
> > > > > > > intel_context *ce)
> > > > > > >
> > > > > > >  static void guc_context_post_unpin(struct intel_context *ce)
> > > > > > >  {
> > > > > > > -   lrc_post_unpin(ce);
> > > > > > > +   __guc_context_post_unpin(ce);
> > > > > > > +}
> > > > > > > +
> > > > > > > +/* Future patches will use this function */
> > > > > > > +__maybe_unused
> > > > > > > +static int guc_parent_context_pre_pin(struct intel_context *ce,
> > > > > > > +                                 struct i915_gem_ww_ctx *ww)
> > > > > > > +{
> > > > > > > +   struct intel_context *child;
> > > > > > > +   int err, i = 0, j = 0;
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           err = i915_active_acquire(&child->active);
> > > > > > > +           if (unlikely(err))
> > > > > > > +                   goto unwind_active;
> > > > > > > +           ++i;
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           err = __guc_context_pre_pin(child, child->engine, ww);
> > > > > > > +           if (unlikely(err))
> > > > > > > +                   goto unwind_pre_pin;
> > > > > > > +           ++j;
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   err = __guc_context_pre_pin(ce, ce->engine, ww);
> > > > > > > +   if (unlikely(err))
> > > > > > > +           goto unwind_pre_pin;
> > > > > > > +
> > > > > > > +   return 0;
> > > > > > > +
> > > > > > > +unwind_pre_pin:
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           if (!j--)
> > > > > > > +                   break;
> > > > > > > +           __guc_context_post_unpin(child);
> > > > > > > +   }
> > > > > > > +
> > > > > > > +unwind_active:
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           if (!i--)
> > > > > > > +                   break;
> > > > > > > +           i915_active_release(&child->active);
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   return err;
> > > > > > > +}
> > > > > > > +
> > > > > > > +/* Future patches will use this function */
> > > > > > > +__maybe_unused
> > > > > > > +static void guc_parent_context_post_unpin(struct intel_context 
> > > > > > > *ce)
> > > > > > > +{
> > > > > > > +   struct intel_context *child;
> > > > > > > +
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           __guc_context_post_unpin(child);
> > > > > > > +   __guc_context_post_unpin(ce);
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           intel_context_get(child);
> > > > > > > +           i915_active_release(&child->active);
> > > > > > > +           intel_context_put(child);
> > > > > > > +   }
> > > > > > > +}
> > > > > > > +
> > > > > > > +/* Future patches will use this function */
> > > > > > > +__maybe_unused
> > > > > > > +static int guc_parent_context_pin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   int ret, i = 0, j = 0;
> > > > > > > +   struct intel_context *child;
> > > > > > > +   struct intel_engine_cs *engine;
> > > > > > > +   intel_engine_mask_t tmp;
> > > > > > > +
> > > > > > > +   GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > > > > +
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           ret = __guc_context_pin(child, child->engine);
> > > > > > > +           if (unlikely(ret))
> > > > > > > +                   goto unwind_pin;
> > > > > > > +           ++i;
> > > > > > > +   }
> > > > > > > +   ret = __guc_context_pin(ce, ce->engine);
> > > > > > > +   if (unlikely(ret))
> > > > > > > +           goto unwind_pin;
> > > > > > > +
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           if (test_bit(CONTEXT_LRCA_DIRTY, &child->flags)) {
> > > > > > > +                   set_bit(CONTEXT_LRCA_DIRTY, &ce->flags);
> > > > > > > +                   break;
> > > > > > > +           }
> > > > > > > +
> > > > > > > +   for_each_engine_masked(engine, ce->engine->gt,
> > > > > > > +                          ce->engine->mask, tmp)
> > > > > > > +           intel_engine_pm_get(engine);
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           for_each_engine_masked(engine, child->engine->gt,
> > > > > > > +                                  child->engine->mask, tmp)
> > > > > > > +                   intel_engine_pm_get(engine);
> > > > > > > +
> > > > > > > +   return 0;
> > > > > > > +
> > > > > > > +unwind_pin:
> > > > > > > +   for_each_child(ce, child) {
> > > > > > > +           if (++j > i)
> > > > > > > +                   break;
> > > > > > > +           __guc_context_unpin(child);
> > > > > > > +   }
> > > > > > > +
> > > > > > > +   return ret;
> > > > > > > +}
> > > > > > > +
> > > > > > > +/* Future patches will use this function */
> > > > > > > +__maybe_unused
> > > > > > > +static void guc_parent_context_unpin(struct intel_context *ce)
> > > > > > > +{
> > > > > > > +   struct intel_context *child;
> > > > > > > +   struct intel_engine_cs *engine;
> > > > > > > +   intel_engine_mask_t tmp;
> > > > > > > +
> > > > > > > +   GEM_BUG_ON(!intel_context_is_parent(ce));
> > > > > > > +   GEM_BUG_ON(context_enabled(ce));
> > > > > > > +
> > > > > > > +   unpin_guc_id(ce_to_guc(ce), ce, true);
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           __guc_context_unpin(child);
> > > > > > > +   __guc_context_unpin(ce);
> > > > > > > +
> > > > > > > +   for_each_engine_masked(engine, ce->engine->gt,
> > > > > > > +                          ce->engine->mask, tmp)
> > > > > > > +           intel_engine_pm_put(engine);
> > > > > > > +   for_each_child(ce, child)
> > > > > > > +           for_each_engine_masked(engine, child->engine->gt,
> > > > > > > +                                  child->engine->mask, tmp)
> > > > > > > +                   intel_engine_pm_put(engine);
> > > > > > >  }
> > > > > > >
> > > > > > >  static void __guc_context_sched_enable(struct intel_guc *guc,
> > > > > > > @@ -2993,18 +3139,17 @@ static int guc_request_alloc(struct 
> > > > > > > i915_request *rq)
> > > > > > >  }
> > > > > > >
> > > > > > >  static int guc_virtual_context_pre_pin(struct intel_context *ce,
> > > > > > > -                                  struct i915_gem_ww_ctx *ww,
> > > > > > > -                                  void **vaddr)
> > > > > > > +                                  struct i915_gem_ww_ctx *ww)
> > > > > > >  {
> > > > > > >     struct intel_engine_cs *engine = 
> > > > > > > guc_virtual_get_sibling(ce->engine, 0);
> > > > > > >
> > > > > > > -   return __guc_context_pre_pin(ce, engine, ww, vaddr);
> > > > > > > +   return __guc_context_pre_pin(ce, engine, ww);
> > > > > > >  }
> > > > > > >
> > > > > > > -static int guc_virtual_context_pin(struct intel_context *ce, 
> > > > > > > void *vaddr)
> > > > > > > +static int guc_virtual_context_pin(struct intel_context *ce)
> > > > > > >  {
> > > > > > >     struct intel_engine_cs *engine = 
> > > > > > > guc_virtual_get_sibling(ce->engine, 0);
> > > > > > > -   int ret = __guc_context_pin(ce, engine, vaddr);
> > > > > > > +   int ret = __guc_context_pin(ce, engine);
> > > > > > >     intel_engine_mask_t tmp, mask = ce->engine->mask;
> > > > > > >
> > > > > > >     if (likely(!ret))
> > > > > > > @@ -3024,7 +3169,7 @@ static void 
> > > > > > > guc_virtual_context_unpin(struct intel_context *ce)
> > > > > > >     GEM_BUG_ON(intel_context_is_barrier(ce));
> > > > > > >
> > > > > > >     unpin_guc_id(guc, ce, true);
> > > > > > > -   lrc_unpin(ce);
> > > > > > > +   __guc_context_unpin(ce);
> > > > > > >
> > > > > > >     for_each_engine_masked(engine, ce->engine->gt, mask, tmp)
> > > > > > >             intel_engine_pm_put(engine);
> > > > > > > --
> > > > > > > 2.28.0
> > > > > > >
> > > > > >
> > > > > > --
> > > > > > Daniel Vetter
> > > > > > Software Engineer, Intel Corporation
> > > > > > http://blog.ffwll.ch
> > > >
> > > > --
> > > > Daniel Vetter
> > > > Software Engineer, Intel Corporation
> > > > http://blog.ffwll.ch
> > >
> > > --
> > > Daniel Vetter
> > > Software Engineer, Intel Corporation
> > > http://blog.ffwll.ch
>
> --
> Daniel Vetter
> Software Engineer, Intel Corporation
> http://blog.ffwll.ch



-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch

Reply via email to