[Intel-gfx] [PATCH v6 2/2] drm/i915/gem: Don't try to map and fence large scanout buffers (v9)

Daniel Vetter daniel at ffwll.ch
Fri Mar 11 09:39:11 UTC 2022


On Mon, 7 Mar 2022 at 21:38, Vivek Kasireddy <vivek.kasireddy at intel.com> wrote:
>
> On platforms capable of allowing 8K (7680 x 4320) modes, pinning 2 or
> more framebuffers/scanout buffers results in only one that is mappable/
> fenceable. Therefore, pageflipping between these 2 FBs where only one
> is mappable/fenceable creates latencies large enough to miss alternate
> vblanks thereby producing less optimal framerate.
>
> This mainly happens because when i915_gem_object_pin_to_display_plane()
> is called to pin one of the FB objs, the associated vma is identified
> as misplaced and therefore i915_vma_unbind() is called which unbinds and
> evicts it. This misplaced vma gets subseqently pinned only when
> i915_gem_object_ggtt_pin_ww() is called without PIN_MAPPABLE. This
> results in a latency of ~10ms and happens every other vblank/repaint cycle.
> Therefore, to fix this issue, we try to see if there is space to map
> at-least two objects of a given size and return early if there isn't. This
> would ensure that we do not try with PIN_MAPPABLE for any objects that
> are too big to map thereby preventing unncessary unbind.
>
> Testcase:
> Running Weston and weston-simple-egl on an Alderlake_S (ADLS) platform
> with a 8K at 60 mode results in only ~40 FPS. Since upstream Weston submits
> a frame ~7ms before the next vblank, the latencies seen between atomic
> commit and flip event are 7, 24 (7 + 16.66), 7, 24..... suggesting that
> it misses the vblank every other frame.
>
> Here is the ftrace snippet that shows the source of the ~10ms latency:
>               i915_gem_object_pin_to_display_plane() {
> 0.102 us   |    i915_gem_object_set_cache_level();
>                 i915_gem_object_ggtt_pin_ww() {
> 0.390 us   |      i915_vma_instance();
> 0.178 us   |      i915_vma_misplaced();
>                   i915_vma_unbind() {
>                   __i915_active_wait() {
> 0.082 us   |        i915_active_acquire_if_busy();
> 0.475 us   |      }
>                   intel_runtime_pm_get() {
> 0.087 us   |        intel_runtime_pm_acquire();
> 0.259 us   |      }
>                   __i915_active_wait() {
> 0.085 us   |        i915_active_acquire_if_busy();
> 0.240 us   |      }
>                   __i915_vma_evict() {
>                     ggtt_unbind_vma() {
>                       gen8_ggtt_clear_range() {
> 10507.255 us |        }
> 10507.689 us |      }
> 10508.516 us |   }
>
> v2: Instead of using bigjoiner checks, determine whether a scanout
>     buffer is too big by checking to see if it is possible to map
>     two of them into the ggtt.
>
> v3 (Ville):
> - Count how many fb objects can be fit into the available holes
>   instead of checking for a hole twice the object size.
> - Take alignment constraints into account.
> - Limit this large scanout buffer check to >= Gen 11 platforms.
>
> v4:
> - Remove existing heuristic that checks just for size. (Ville)
> - Return early if we find space to map at-least two objects. (Tvrtko)
> - Slightly update the commit message.
>
> v5: (Tvrtko)
> - Rename the function to indicate that the object may be too big to
>   map into the aperture.
> - Account for guard pages while calculating the total size required
>   for the object.
> - Do not subject all objects to the heuristic check and instead
>   consider objects only of a certain size.
> - Do the hole walk using the rbtree.
> - Preserve the existing PIN_NONBLOCK logic.
> - Drop the PIN_MAPPABLE check while pinning the VMA.
>
> v6: (Tvrtko)
> - Return 0 on success and the specific error code on failure to
>   preserve the existing behavior.
>
> v7: (Ville)
> - Drop the HAS_GMCH(i915), DISPLAY_VER(i915) < 11 and
>   size < ggtt->mappable_end / 4 checks.
> - Drop the redundant check that is based on previous heuristic.
>
> v8:
> - Make sure that we are holding the mutex associated with ggtt vm
>   as we traverse the hole nodes.
>
> v9: (Tvrtko)
> - Use mutex_lock_interruptible_nested() instead of mutex_lock().
>
> Cc: Ville Syrjälä <ville.syrjala at linux.intel.com>
> Cc: Maarten Lankhorst <maarten.lankhorst at linux.intel.com>
> Cc: Tvrtko Ursulin <tvrtko.ursulin at linux.intel.com>
> Cc: Manasi Navare <manasi.d.navare at intel.com>
> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
> Signed-off-by: Vivek Kasireddy <vivek.kasireddy at intel.com>
> ---
>  drivers/gpu/drm/i915/i915_gem.c | 128 +++++++++++++++++++++++---------
>  1 file changed, 94 insertions(+), 34 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 9747924cc57b..e0d731b3f215 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -49,6 +49,7 @@
>  #include "gem/i915_gem_pm.h"
>  #include "gem/i915_gem_region.h"
>  #include "gem/i915_gem_userptr.h"
> +#include "gem/i915_gem_tiling.h"
>  #include "gt/intel_engine_user.h"
>  #include "gt/intel_gt.h"
>  #include "gt/intel_gt_pm.h"
> @@ -882,6 +883,96 @@ static void discard_ggtt_vma(struct i915_vma *vma)
>         spin_unlock(&obj->vma.lock);
>  }
>
> +static int
> +i915_gem_object_fits_in_aperture(struct drm_i915_gem_object *obj,
> +                                u64 alignment, u64 flags)

Tvrtko asked me to ack the first patch, but then I looked at this and
started wondering.

Conceptually this doesn't pass the smell test. What if we have
multiple per-crtc buffers? Multiple planes on the same crtc? What if
the app does triple buffer? You'll be forever busy tuning this
heuristics, which can't fundamentally be fixed I think. The old "half
of mappable" heuristic isn't really better, but at least it was dead
simple.

Imo what we need here is a change in approach:
1. Check whether the useable view for scanout exists already. If yes,
use that. This should avoid the constant unbinding stalls.
2. Try to in buffer to mappabley, but without evicting anything (so
not the non-blocking thing)
3. Pin the buffer with the most lenient approach

Even the non-blocking interim stage is dangerous, since it'll just
result in other buffers (e.g. when triple-buffering) getting unbound
and we're back to the same stall. Note that this could have an impact
on cpu rendering compositors, where we might end up relying a lot more
partial views. But as long as we are a tad more aggressive (i.e. the
non-blocking binding) in the mmap path that should work out to keep
everything balanced, since usually you render first before you display
anything. And so the buffer should end up in the ideal place.

I'd try to first skip the 2. step since I think it'll require a bit of
work, and frankly I don't think we care about the potential fallout.
-Daniel

> +{
> +       struct drm_i915_private *i915 = to_i915(obj->base.dev);
> +       struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
> +       struct drm_mm_node *hole;
> +       u64 hole_start, hole_end, start, end;
> +       u64 fence_size, fence_alignment;
> +       unsigned int count = 0;
> +       int err = 0;
> +
> +       /*
> +        * If the required space is larger than the available
> +        * aperture, we will not able to find a slot for the
> +        * object and unbinding the object now will be in
> +        * vain. Worse, doing so may cause us to ping-pong
> +        * the object in and out of the Global GTT and
> +        * waste a lot of cycles under the mutex.
> +        */
> +       if (obj->base.size > ggtt->mappable_end)
> +               return -E2BIG;
> +
> +       /*
> +        * If NONBLOCK is set the caller is optimistically
> +        * trying to cache the full object within the mappable
> +        * aperture, and *must* have a fallback in place for
> +        * situations where we cannot bind the object. We
> +        * can be a little more lax here and use the fallback
> +        * more often to avoid costly migrations of ourselves
> +        * and other objects within the aperture.
> +        */
> +       if (!(flags & PIN_NONBLOCK))
> +               return 0;
> +
> +       /*
> +        * Other objects such as batchbuffers are fairly small compared
> +        * to FBs and are unlikely to exahust the aperture space.
> +        * Therefore, return early if this obj is not an FB.
> +        */
> +       if (!i915_gem_object_is_framebuffer(obj))
> +               return 0;
> +
> +       fence_size = i915_gem_fence_size(i915, obj->base.size,
> +                                        i915_gem_object_get_tiling(obj),
> +                                        i915_gem_object_get_stride(obj));
> +
> +       if (i915_vm_has_cache_coloring(&ggtt->vm))
> +               fence_size += 2 * I915_GTT_PAGE_SIZE;
> +
> +       fence_alignment = i915_gem_fence_alignment(i915, obj->base.size,
> +                                                  i915_gem_object_get_tiling(obj),
> +                                                  i915_gem_object_get_stride(obj));
> +       alignment = max_t(u64, alignment, fence_alignment);
> +
> +       err = mutex_lock_interruptible_nested(&ggtt->vm.mutex, 0);
> +       if (err)
> +               return err;
> +
> +       /*
> +        * Assuming this object is a large scanout buffer, we try to find
> +        * out if there is room to map at-least two of them. There could
> +        * be space available to map one but to be consistent, we try to
> +        * avoid mapping/fencing any of them.
> +        */
> +       drm_mm_for_each_suitable_hole(hole, &ggtt->vm.mm, 0, ggtt->mappable_end,
> +                                     fence_size, DRM_MM_INSERT_LOW) {
> +               hole_start = drm_mm_hole_node_start(hole);
> +               hole_end = hole_start + hole->hole_size;
> +
> +               do {
> +                       start = round_up(hole_start, alignment);
> +                       end = min_t(u64, hole_end, ggtt->mappable_end);
> +
> +                       if (range_overflows(start, fence_size, end))
> +                               break;
> +
> +                       if (++count >= 2) {
> +                               mutex_unlock(&ggtt->vm.mutex);
> +                               return 0;
> +                       }
> +
> +                       hole_start = start + fence_size;
> +               } while (1);
> +       }
> +
> +       mutex_unlock(&ggtt->vm.mutex);
> +       return -ENOSPC;
> +}
> +
>  struct i915_vma *
>  i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
>                             struct i915_gem_ww_ctx *ww,
> @@ -897,36 +988,9 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
>
>         if (flags & PIN_MAPPABLE &&
>             (!view || view->type == I915_GGTT_VIEW_NORMAL)) {
> -               /*
> -                * If the required space is larger than the available
> -                * aperture, we will not able to find a slot for the
> -                * object and unbinding the object now will be in
> -                * vain. Worse, doing so may cause us to ping-pong
> -                * the object in and out of the Global GTT and
> -                * waste a lot of cycles under the mutex.
> -                */
> -               if (obj->base.size > ggtt->mappable_end)
> -                       return ERR_PTR(-E2BIG);
> -
> -               /*
> -                * If NONBLOCK is set the caller is optimistically
> -                * trying to cache the full object within the mappable
> -                * aperture, and *must* have a fallback in place for
> -                * situations where we cannot bind the object. We
> -                * can be a little more lax here and use the fallback
> -                * more often to avoid costly migrations of ourselves
> -                * and other objects within the aperture.
> -                *
> -                * Half-the-aperture is used as a simple heuristic.
> -                * More interesting would to do search for a free
> -                * block prior to making the commitment to unbind.
> -                * That caters for the self-harm case, and with a
> -                * little more heuristics (e.g. NOFAULT, NOEVICT)
> -                * we could try to minimise harm to others.
> -                */
> -               if (flags & PIN_NONBLOCK &&
> -                   obj->base.size > ggtt->mappable_end / 2)
> -                       return ERR_PTR(-ENOSPC);
> +               ret = i915_gem_object_fits_in_aperture(obj, alignment, flags);
> +               if (ret)
> +                       return ERR_PTR(ret);
>         }
>
>  new_vma:
> @@ -938,10 +1002,6 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj,
>                 if (flags & PIN_NONBLOCK) {
>                         if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma))
>                                 return ERR_PTR(-ENOSPC);
> -
> -                       if (flags & PIN_MAPPABLE &&
> -                           vma->fence_size > ggtt->mappable_end / 2)
> -                               return ERR_PTR(-ENOSPC);
>                 }
>
>                 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) {
> --
> 2.35.1
>


-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch


More information about the Intel-gfx mailing list