[Intel-gfx] [PATCH 28/28] drm/i915: Remove short-term pins from execbuf, v4.

Maarten Lankhorst maarten.lankhorst at linux.intel.com
Mon Nov 29 13:44:11 UTC 2021


On 25-10-2021 17:02, Matthew Auld wrote:
> On Thu, 21 Oct 2021 at 11:37, Maarten Lankhorst
> <maarten.lankhorst at linux.intel.com> wrote:
>> Add a flag PIN_VALIDATE, to indicate we don't need to pin and only
>> protected by the object lock.
>>
>> This removes the need to unpin, which is done by just releasing the
>> lock.
>>
>> eb_reserve is slightly reworked for readability, but the same steps
>> are still done:
>> - First pass pins with NONBLOCK.
>> - Second pass unbinds all objects first, then pins.
>> - Third pass is only called when not all objects are softpinned, and
>>   unbinds all objects, then calls i915_gem_evict_vm(), then pins.
>>
>> When evicting the entire vm in eb_reserve() we do temporarily pin objects
>> that are marked with EXEC_OBJECT_PINNED. This is because they are already
>> at their destination, and i915_gem_evict_vm() would otherwise unbind them.
>>
>> However, we reduce the visibility of those pins by limiting the pin
>> to our call to i915_gem_evict_vm() only, and pin with vm->mutex held,
>> instead of the entire duration of the execbuf.
>>
>> Not sure the latter matters, one can hope..
>> In theory we could kill the pinning by adding an extra flag to the vma
>> to temporarily prevent unbinding for gtt for i915_gem_evict_vm only, but
>> I think that might be overkill. We're still holding the object lock, and
>> we don't have blocking eviction yet. It's likely sufficient to simply
>> enforce EXEC_OBJECT_PINNED for all objects on >= gen12.
>>
>> Changes since v1:
>> - Split out eb_reserve() into separate functions for readability.
>> Changes since v2:
>> - Make batch buffer mappable on platforms where only GGTT is available,
>>   to prevent moving the batch buffer during relocations.
>> Changes since v3:
>> - Preserve current behavior for batch buffer, instead be cautious when
>>   calling i915_gem_object_ggtt_pin_ww, and re-use the current batch vma
>>   if it's inside ggtt and map-and-fenceable.
>>
>> Signed-off-by: Maarten Lankhorst <maarten.lankhorst at linux.intel.com>
>> ---
>>  .../gpu/drm/i915/gem/i915_gem_execbuffer.c    | 252 ++++++++++--------
>>  drivers/gpu/drm/i915/i915_gem_gtt.h           |   1 +
>>  drivers/gpu/drm/i915/i915_vma.c               |  24 +-
>>  3 files changed, 161 insertions(+), 116 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>> index bbf2a10738f7..19f91143cfcf 100644
>> --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>> +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>> @@ -439,7 +439,7 @@ eb_pin_vma(struct i915_execbuffer *eb,
>>         else
>>                 pin_flags = entry->offset & PIN_OFFSET_MASK;
>>
>> -       pin_flags |= PIN_USER | PIN_NOEVICT | PIN_OFFSET_FIXED;
>> +       pin_flags |= PIN_USER | PIN_NOEVICT | PIN_OFFSET_FIXED | PIN_VALIDATE;
>>         if (unlikely(ev->flags & EXEC_OBJECT_NEEDS_GTT))
>>                 pin_flags |= PIN_GLOBAL;
>>
>> @@ -457,17 +457,15 @@ eb_pin_vma(struct i915_execbuffer *eb,
>>                                              entry->pad_to_size,
>>                                              entry->alignment,
>>                                              eb_pin_flags(entry, ev->flags) |
>> -                                            PIN_USER | PIN_NOEVICT);
>> +                                            PIN_USER | PIN_NOEVICT | PIN_VALIDATE);
>>                 if (unlikely(err))
>>                         return err;
>>         }
>>
>>         if (unlikely(ev->flags & EXEC_OBJECT_NEEDS_FENCE)) {
>>                 err = i915_vma_pin_fence(vma);
>> -               if (unlikely(err)) {
>> -                       i915_vma_unpin(vma);
>> +               if (unlikely(err))
>>                         return err;
>> -               }
>>
>>                 if (vma->fence)
>>                         ev->flags |= __EXEC_OBJECT_HAS_FENCE;
>> @@ -483,13 +481,9 @@ eb_pin_vma(struct i915_execbuffer *eb,
>>  static inline void
>>  eb_unreserve_vma(struct eb_vma *ev)
>>  {
>> -       if (!(ev->flags & __EXEC_OBJECT_HAS_PIN))
>> -               return;
>> -
>>         if (unlikely(ev->flags & __EXEC_OBJECT_HAS_FENCE))
>>                 __i915_vma_unpin_fence(ev->vma);
>>
>> -       __i915_vma_unpin(ev->vma);
>>         ev->flags &= ~__EXEC_OBJECT_RESERVED;
>>  }
>>
>> @@ -682,10 +676,8 @@ static int eb_reserve_vma(struct i915_execbuffer *eb,
>>
>>         if (unlikely(ev->flags & EXEC_OBJECT_NEEDS_FENCE)) {
>>                 err = i915_vma_pin_fence(vma);
>> -               if (unlikely(err)) {
>> -                       i915_vma_unpin(vma);
>> +               if (unlikely(err))
>>                         return err;
>> -               }
>>
>>                 if (vma->fence)
>>                         ev->flags |= __EXEC_OBJECT_HAS_FENCE;
>> @@ -697,85 +689,129 @@ static int eb_reserve_vma(struct i915_execbuffer *eb,
>>         return 0;
>>  }
>>
>> -static int eb_reserve(struct i915_execbuffer *eb)
>> +static int eb_evict_vm(struct i915_execbuffer *eb)
>> +{
>> +       const unsigned int count = eb->buffer_count;
>> +       unsigned int i;
>> +       int err;
>> +
>> +       err = mutex_lock_interruptible(&eb->context->vm->mutex);
>> +       if (err)
>> +               return err;
>> +
>> +       /* pin to protect against i915_gem_evict_vm evicting below */
>> +       for (i = 0; i < count; i++) {
>> +               struct eb_vma *ev = &eb->vma[i];
>> +
>> +               if (ev->flags & __EXEC_OBJECT_HAS_PIN)
>> +                       __i915_vma_pin(ev->vma);
>> +       }
>> +
>> +       /* Too fragmented, unbind everything and retry */
>> +       err = i915_gem_evict_vm(eb->context->vm, &eb->ww);
>> +
>> +       /* unpin objects.. */
>> +       for (i = 0; i < count; i++) {
>> +               struct eb_vma *ev = &eb->vma[i];
>> +
>> +               if (ev->flags & __EXEC_OBJECT_HAS_PIN)
>> +                       i915_vma_unpin(ev->vma);
>> +       }
>> +
>> +       mutex_unlock(&eb->context->vm->mutex);
>> +
>> +       return err;
>> +}
>> +
>> +static bool eb_unbind(struct i915_execbuffer *eb)
>>  {
>>         const unsigned int count = eb->buffer_count;
>> -       unsigned int pin_flags = PIN_USER | PIN_NONBLOCK;
>> +       unsigned int i;
>>         struct list_head last;
>> +       bool unpinned = false;
>> +
>> +       /* Resort *all* the objects into priority order */
>> +       INIT_LIST_HEAD(&eb->unbound);
>> +       INIT_LIST_HEAD(&last);
>> +
>> +       for (i = 0; i < count; i++) {
>> +               struct eb_vma *ev = &eb->vma[i];
>> +               unsigned int flags = ev->flags;
>> +
>> +               if (flags & EXEC_OBJECT_PINNED &&
>> +                   flags & __EXEC_OBJECT_HAS_PIN)
>> +                       continue;
>> +
>> +               unpinned = true;
>> +               eb_unreserve_vma(ev);
>> +
>> +               if (flags & EXEC_OBJECT_PINNED)
>> +                       /* Pinned must have their slot */
>> +                       list_add(&ev->bind_link, &eb->unbound);
>> +               else if (flags & __EXEC_OBJECT_NEEDS_MAP)
>> +                       /* Map require the lowest 256MiB (aperture) */
>> +                       list_add_tail(&ev->bind_link, &eb->unbound);
>> +               else if (!(flags & EXEC_OBJECT_SUPPORTS_48B_ADDRESS))
>> +                       /* Prioritise 4GiB region for restricted bo */
>> +                       list_add(&ev->bind_link, &last);
>> +               else
>> +                       list_add_tail(&ev->bind_link, &last);
>> +       }
>> +
>> +       list_splice_tail(&last, &eb->unbound);
>> +       return unpinned;
>> +}
>> +
>> +static int eb_reserve(struct i915_execbuffer *eb)
>> +{
>>         struct eb_vma *ev;
>> -       unsigned int i, pass;
>> +       unsigned int pass;
>>         int err = 0;
>> +       bool unpinned;
>>
>>         /*
>>          * Attempt to pin all of the buffers into the GTT.
>> -        * This is done in 3 phases:
>> +        * This is done in 2 phases:
>>          *
>> -        * 1a. Unbind all objects that do not match the GTT constraints for
>> -        *     the execbuffer (fenceable, mappable, alignment etc).
>> -        * 1b. Increment pin count for already bound objects.
>> -        * 2.  Bind new objects.
>> -        * 3.  Decrement pin count.
>> +        * 1. Unbind all objects that do not match the GTT constraints for
>> +        *    the execbuffer (fenceable, mappable, alignment etc).
>> +        * 2. Bind new objects.
>>          *
>>          * This avoid unnecessary unbinding of later objects in order to make
>>          * room for the earlier objects *unless* we need to defragment.
>> +        *
>> +        * Defragmenting is skipped if all objects are pinned at a fixed location.
>>          */
>> -       pass = 0;
>> -       do {
>> -               list_for_each_entry(ev, &eb->unbound, bind_link) {
>> -                       err = eb_reserve_vma(eb, ev, pin_flags);
>> -                       if (err)
>> -                               break;
>> -               }
>> -               if (err != -ENOSPC)
>> -                       return err;
>> +       for (pass = 0; pass <= 2; pass++) {
>> +               int pin_flags = PIN_USER | PIN_VALIDATE;
>>
>> -               /* Resort *all* the objects into priority order */
>> -               INIT_LIST_HEAD(&eb->unbound);
>> -               INIT_LIST_HEAD(&last);
>> -               for (i = 0; i < count; i++) {
>> -                       unsigned int flags;
>> +               if (pass == 0)
>> +                       pin_flags |= PIN_NONBLOCK;
>>
>> -                       ev = &eb->vma[i];
>> -                       flags = ev->flags;
>> -                       if (flags & EXEC_OBJECT_PINNED &&
>> -                           flags & __EXEC_OBJECT_HAS_PIN)
>> -                               continue;
>> +               if (pass >= 1)
>> +                       unpinned = eb_unbind(eb);
>>
>> -                       eb_unreserve_vma(ev);
>> -
>> -                       if (flags & EXEC_OBJECT_PINNED)
>> -                               /* Pinned must have their slot */
>> -                               list_add(&ev->bind_link, &eb->unbound);
>> -                       else if (flags & __EXEC_OBJECT_NEEDS_MAP)
>> -                               /* Map require the lowest 256MiB (aperture) */
>> -                               list_add_tail(&ev->bind_link, &eb->unbound);
>> -                       else if (!(flags & EXEC_OBJECT_SUPPORTS_48B_ADDRESS))
>> -                               /* Prioritise 4GiB region for restricted bo */
>> -                               list_add(&ev->bind_link, &last);
>> -                       else
>> -                               list_add_tail(&ev->bind_link, &last);
>> -               }
>> -               list_splice_tail(&last, &eb->unbound);
>> -
>> -               switch (pass++) {
>> -               case 0:
>> -                       break;
>> +               if (pass == 2) {
>> +                       /* No point in defragmenting gtt if all is pinned */
>> +                       if (!unpinned)
>> +                               return -ENOSPC;
> Can this ever happen? If everything is already pinned where it's meant
> to be, then how did we get here?
Hmm no, in previous versions it could, but it looks like I removed that in a refactor. Seems like dead code.
>
>> -               case 1:
>> -                       /* Too fragmented, unbind everything and retry */
>> -                       mutex_lock(&eb->context->vm->mutex);
>> -                       err = i915_gem_evict_vm(eb->context->vm, &eb->ww);
>> -                       mutex_unlock(&eb->context->vm->mutex);
>> +                       err = eb_evict_vm(eb);
>>                         if (err)
>>                                 return err;
>> -                       break;
>> +               }
>>
>> -               default:
>> -                       return -ENOSPC;
>> +               list_for_each_entry(ev, &eb->unbound, bind_link) {
>> +                       err = eb_reserve_vma(eb, ev, pin_flags);
>> +                       if (err)
>> +                               break;
>>                 }
>>
>> -               pin_flags = PIN_USER;
>> -       } while (1);
>> +               if (err != -ENOSPC)
>> +                       break;
>> +       }
>> +
>> +       return err;
>>  }
>>
>>  static int eb_select_context(struct i915_execbuffer *eb)
>> @@ -1184,10 +1220,11 @@ static void *reloc_kmap(struct drm_i915_gem_object *obj,
>>         return vaddr;
>>  }
>>
>> -static void *reloc_iomap(struct drm_i915_gem_object *obj,
>> +static void *reloc_iomap(struct i915_vma *batch,
>>                          struct i915_execbuffer *eb,
>>                          unsigned long page)
>>  {
>> +       struct drm_i915_gem_object *obj = batch->obj;
>>         struct reloc_cache *cache = &eb->reloc_cache;
>>         struct i915_ggtt *ggtt = cache_to_ggtt(cache);
>>         unsigned long offset;
>> @@ -1197,7 +1234,7 @@ static void *reloc_iomap(struct drm_i915_gem_object *obj,
>>                 intel_gt_flush_ggtt_writes(ggtt->vm.gt);
>>                 io_mapping_unmap_atomic((void __force __iomem *) unmask_page(cache->vaddr));
>>         } else {
>> -               struct i915_vma *vma;
>> +               struct i915_vma *vma = ERR_PTR(-ENODEV);
>>                 int err;
>>
>>                 if (i915_gem_object_is_tiled(obj))
>> @@ -1210,10 +1247,23 @@ static void *reloc_iomap(struct drm_i915_gem_object *obj,
>>                 if (err)
>>                         return ERR_PTR(err);
>>
>> -               vma = i915_gem_object_ggtt_pin_ww(obj, &eb->ww, NULL, 0, 0,
>> -                                                 PIN_MAPPABLE |
>> -                                                 PIN_NONBLOCK /* NOWARN */ |
>> -                                                 PIN_NOEVICT);
>> +               /*
>> +                * i915_gem_object_ggtt_pin_ww may attempt to remove the batch
>> +                * VMA from the object list because we no longer pin.
>> +                *
>> +                * Only attempt to pin the batch buffer to ggtt if the current batch
>> +                * is not inside ggtt, or the batch buffer is not misplaced.
>> +                */
>> +               if (!i915_is_ggtt(batch->vm)) {
>> +                       vma = i915_gem_object_ggtt_pin_ww(obj, &eb->ww, NULL, 0, 0,
>> +                                                         PIN_MAPPABLE |
>> +                                                         PIN_NONBLOCK /* NOWARN */ |
>> +                                                         PIN_NOEVICT);
>> +               } else if (i915_vma_is_map_and_fenceable(batch)) {
>> +                       __i915_vma_pin(batch);
>> +                       vma = batch;
>> +               }
>> +
>>                 if (vma == ERR_PTR(-EDEADLK))
>>                         return vma;
>>
>> @@ -1251,7 +1301,7 @@ static void *reloc_iomap(struct drm_i915_gem_object *obj,
>>         return vaddr;
>>  }
>>
>> -static void *reloc_vaddr(struct drm_i915_gem_object *obj,
>> +static void *reloc_vaddr(struct i915_vma *vma,
>>                          struct i915_execbuffer *eb,
>>                          unsigned long page)
>>  {
>> @@ -1263,9 +1313,9 @@ static void *reloc_vaddr(struct drm_i915_gem_object *obj,
>>         } else {
>>                 vaddr = NULL;
>>                 if ((cache->vaddr & KMAP) == 0)
>> -                       vaddr = reloc_iomap(obj, eb, page);
>> +                       vaddr = reloc_iomap(vma, eb, page);
>>                 if (!vaddr)
>> -                       vaddr = reloc_kmap(obj, cache, page);
>> +                       vaddr = reloc_kmap(vma->obj, cache, page);
>>         }
>>
>>         return vaddr;
>> @@ -1306,7 +1356,7 @@ relocate_entry(struct i915_vma *vma,
>>         void *vaddr;
>>
>>  repeat:
>> -       vaddr = reloc_vaddr(vma->obj, eb,
>> +       vaddr = reloc_vaddr(vma, eb,
>>                             offset >> PAGE_SHIFT);
>>         if (IS_ERR(vaddr))
>>                 return PTR_ERR(vaddr);
>> @@ -2074,7 +2124,7 @@ shadow_batch_pin(struct i915_execbuffer *eb,
>>         if (IS_ERR(vma))
>>                 return vma;
>>
>> -       err = i915_vma_pin_ww(vma, &eb->ww, 0, 0, flags);
>> +       err = i915_vma_pin_ww(vma, &eb->ww, 0, 0, flags | PIN_VALIDATE);
>>         if (err)
>>                 return ERR_PTR(err);
>>
>> @@ -2088,7 +2138,7 @@ static struct i915_vma *eb_dispatch_secure(struct i915_execbuffer *eb, struct i9
>>          * batch" bit. Hence we need to pin secure batches into the global gtt.
>>          * hsw should have this fixed, but bdw mucks it up again. */
>>         if (eb->batch_flags & I915_DISPATCH_SECURE)
>> -               return i915_gem_object_ggtt_pin_ww(vma->obj, &eb->ww, NULL, 0, 0, 0);
>> +               return i915_gem_object_ggtt_pin_ww(vma->obj, &eb->ww, NULL, 0, 0, PIN_VALIDATE);
>>
>>         return NULL;
>>  }
>> @@ -2139,13 +2189,12 @@ static int eb_parse(struct i915_execbuffer *eb)
>>
>>         err = i915_gem_object_lock(pool->obj, &eb->ww);
>>         if (err)
>> -               goto err;
>> +               return err;
>>
>>         shadow = shadow_batch_pin(eb, pool->obj, eb->context->vm, PIN_USER);
>> -       if (IS_ERR(shadow)) {
>> -               err = PTR_ERR(shadow);
>> -               goto err;
>> -       }
>> +       if (IS_ERR(shadow))
>> +               return PTR_ERR(shadow);
>> +
>>         intel_gt_buffer_pool_mark_used(pool);
>>         i915_gem_object_set_readonly(shadow->obj);
>>         shadow->private = pool;
>> @@ -2157,25 +2206,21 @@ static int eb_parse(struct i915_execbuffer *eb)
>>                 shadow = shadow_batch_pin(eb, pool->obj,
>>                                           &eb->gt->ggtt->vm,
>>                                           PIN_GLOBAL);
>> -               if (IS_ERR(shadow)) {
>> -                       err = PTR_ERR(shadow);
>> -                       shadow = trampoline;
>> -                       goto err_shadow;
>> -               }
>> +               if (IS_ERR(shadow))
>> +                       return PTR_ERR(shadow);
>> +
>>                 shadow->private = pool;
>>
>>                 eb->batch_flags |= I915_DISPATCH_SECURE;
>>         }
>>
>>         batch = eb_dispatch_secure(eb, shadow);
>> -       if (IS_ERR(batch)) {
>> -               err = PTR_ERR(batch);
>> -               goto err_trampoline;
>> -       }
>> +       if (IS_ERR(batch))
>> +               return PTR_ERR(batch);
>>
>>         err = dma_resv_reserve_shared(shadow->obj->base.resv, 1);
>>         if (err)
>> -               goto err_trampoline;
>> +               return err;
>>
>>         err = intel_engine_cmd_parser(eb->context->engine,
>>                                       eb->batches[0]->vma,
>> @@ -2183,7 +2228,7 @@ static int eb_parse(struct i915_execbuffer *eb)
>>                                       eb->batch_len[0],
>>                                       shadow, trampoline);
>>         if (err)
>> -               goto err_unpin_batch;
>> +               return err;
>>
>>         eb->batches[0] = &eb->vma[eb->buffer_count++];
>>         eb->batches[0]->vma = i915_vma_get(shadow);
>> @@ -2202,17 +2247,6 @@ static int eb_parse(struct i915_execbuffer *eb)
>>                 eb->batches[0]->vma = i915_vma_get(batch);
>>         }
>>         return 0;
>> -
>> -err_unpin_batch:
>> -       if (batch)
>> -               i915_vma_unpin(batch);
>> -err_trampoline:
>> -       if (trampoline)
>> -               i915_vma_unpin(trampoline);
>> -err_shadow:
>> -       i915_vma_unpin(shadow);
>> -err:
>> -       return err;
>>  }
>>
>>  static int eb_request_submit(struct i915_execbuffer *eb,
>> @@ -3337,8 +3371,6 @@ i915_gem_do_execbuffer(struct drm_device *dev,
>>
>>  err_vma:
>>         eb_release_vmas(&eb, true);
>> -       if (eb.trampoline)
>> -               i915_vma_unpin(eb.trampoline);
>>         WARN_ON(err == -EDEADLK);
>>         i915_gem_ww_ctx_fini(&eb.ww);
>>
>> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
>> index e4938aba3fe9..8c2f57eb5dda 100644
>> --- a/drivers/gpu/drm/i915/i915_gem_gtt.h
>> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
>> @@ -44,6 +44,7 @@ int i915_gem_gtt_insert(struct i915_address_space *vm,
>>  #define PIN_HIGH               BIT_ULL(5)
>>  #define PIN_OFFSET_BIAS                BIT_ULL(6)
>>  #define PIN_OFFSET_FIXED       BIT_ULL(7)
>> +#define PIN_VALIDATE           BIT_ULL(8) /* validate placement only, no need to call unpin() */
>>
>>  #define PIN_GLOBAL             BIT_ULL(10) /* I915_VMA_GLOBAL_BIND */
>>  #define PIN_USER               BIT_ULL(11) /* I915_VMA_LOCAL_BIND */
>> diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
>> index 65168db534f0..0706731b211d 100644
>> --- a/drivers/gpu/drm/i915/i915_vma.c
>> +++ b/drivers/gpu/drm/i915/i915_vma.c
>> @@ -751,6 +751,15 @@ static bool try_qad_pin(struct i915_vma *vma, unsigned int flags)
>>         unsigned int bound;
>>
>>         bound = atomic_read(&vma->flags);
>> +
>> +       if (flags & PIN_VALIDATE) {
>> +               flags &= I915_VMA_BIND_MASK;
>> +
>> +               return (flags & bound) == flags;
>> +       }
>> +
>> +       /* with the lock mandatory for unbind, we don't race here */
>> +       flags &= I915_VMA_BIND_MASK;
>>         do {
>>                 if (unlikely(flags & ~bound))
>>                         return false;
>> @@ -1176,7 +1185,7 @@ int i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
>>         GEM_BUG_ON(!(flags & (PIN_USER | PIN_GLOBAL)));
>>
>>         /* First try and grab the pin without rebinding the vma */
>> -       if (try_qad_pin(vma, flags & I915_VMA_BIND_MASK))
>> +       if (try_qad_pin(vma, flags))
>>                 return 0;
>>
>>         err = i915_vma_get_pages(vma);
>> @@ -1255,7 +1264,8 @@ int i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
>>         }
>>
>>         if (unlikely(!(flags & ~bound & I915_VMA_BIND_MASK))) {
>> -               __i915_vma_pin(vma);
>> +               if (!(flags & PIN_VALIDATE))
>> +                       __i915_vma_pin(vma);
>>                 goto err_unlock;
>>         }
>>
>> @@ -1284,8 +1294,10 @@ int i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
>>         atomic_add(I915_VMA_PAGES_ACTIVE, &vma->pages_count);
>>         list_move_tail(&vma->vm_link, &vma->vm->bound_list);
>>
>> -       __i915_vma_pin(vma);
>> -       GEM_BUG_ON(!i915_vma_is_pinned(vma));
>> +       if (!(flags & PIN_VALIDATE)) {
>> +               __i915_vma_pin(vma);
>> +               GEM_BUG_ON(!i915_vma_is_pinned(vma));
>> +       }
>>         GEM_BUG_ON(!i915_vma_is_bound(vma, flags));
>>         GEM_BUG_ON(i915_vma_misplaced(vma, size, alignment, flags));
>>
>> @@ -1538,8 +1550,6 @@ static int __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *
>>  {
>>         int err;
>>
>> -       GEM_BUG_ON(!i915_vma_is_pinned(vma));
>> -
>>         /* Wait for the vma to be bound before we start! */
>>         err = __i915_request_await_bind(rq, vma);
>>         if (err)
>> @@ -1558,6 +1568,8 @@ int _i915_vma_move_to_active(struct i915_vma *vma,
>>
>>         assert_object_held(obj);
>>
>> +       GEM_BUG_ON(!vma->pages);
>> +
>>         err = __i915_vma_move_to_active(vma, rq);
>>         if (unlikely(err))
>>                 return err;
>> --
>> 2.33.0
>>



More information about the Intel-gfx mailing list