[Intel-gfx] [PATCH 1/3] drm: Track drm_mm nodes with an interval tree

David Herrmann dh.herrmann at gmail.com
Wed Aug 3 17:55:38 UTC 2016


Hey

On Wed, Aug 3, 2016 at 5:04 PM, Chris Wilson <chris at chris-wilson.co.uk> wrote:
> In addition to the last-in/first-out stack for accessing drm_mm nodes,
> we occasionally and in the future often want to find a drm_mm_node by an
> address. To do so efficiently we need to track the nodes in an interval
> tree - lookups for a particular address will then be O(lg(N)), where N
> is the number of nodes in the range manager as opposed to O(N).
> Insertion however gains an extra O(lg(N)) step for all nodes
> irrespective of whether the interval tree is in use. For future i915
> patches, eliminating the linear walk is a significant improvement.
>
> v2: Use generic interval-tree template for u64 and faster insertion.
>
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> Cc: David Herrmann <dh.herrmann at gmail.com>
> Cc: dri-devel at lists.freedesktop.org
> ---
>  drivers/gpu/drm/drm_mm.c | 133 +++++++++++++++++++++++++++++++++++++++--------
>  include/drm/drm_mm.h     |  12 +++++
>  2 files changed, 122 insertions(+), 23 deletions(-)
>
> diff --git a/drivers/gpu/drm/drm_mm.c b/drivers/gpu/drm/drm_mm.c
> index cb39f45d6a16..5c188c56894b 100644
> --- a/drivers/gpu/drm/drm_mm.c
> +++ b/drivers/gpu/drm/drm_mm.c
> @@ -46,6 +46,7 @@
>  #include <linux/slab.h>
>  #include <linux/seq_file.h>
>  #include <linux/export.h>
> +#include <linux/interval_tree_generic.h>
>
>  /**
>   * DOC: Overview
> @@ -103,6 +104,72 @@ static struct drm_mm_node *drm_mm_search_free_in_range_generic(const struct drm_
>                                                 u64 end,
>                                                 enum drm_mm_search_flags flags);
>
> +#define START(node) ((node)->start)
> +#define LAST(node)  ((node)->start + (node)->size - 1)

So this goes nuts with "size == 0". We do not explicitly prevent that
from happening, I think, but might be prevented in the upper layers.
Might wanna add WARN_ONs?

Otherwise, looks good to me:

Reviewed-by: David Herrmann <dh.herrmann at gmail.com>

Thanks
David

> +
> +INTERVAL_TREE_DEFINE(struct drm_mm_node, rb,
> +                    u64, __subtree_last,
> +                    START, LAST, static inline, drm_mm_interval_tree)
> +
> +struct drm_mm_node *
> +drm_mm_interval_first(struct drm_mm *mm, u64 start, u64 last)
> +{
> +       return drm_mm_interval_tree_iter_first(&mm->interval_tree,
> +                                              start, last);
> +}
> +EXPORT_SYMBOL(drm_mm_interval_first);
> +
> +struct drm_mm_node *
> +drm_mm_interval_next(struct drm_mm_node *node, u64 start, u64 last)
> +{
> +       return drm_mm_interval_tree_iter_next(node, start, last);
> +}
> +EXPORT_SYMBOL(drm_mm_interval_next);
> +
> +static void drm_mm_interval_tree_add_node(struct drm_mm_node *hole_node,
> +                                         struct drm_mm_node *node)
> +{
> +       struct drm_mm *mm = hole_node->mm;
> +       struct rb_node **link, *rb;
> +       struct drm_mm_node *parent;
> +
> +       node->__subtree_last = LAST(node);
> +
> +       if (hole_node->allocated) {
> +               rb = &hole_node->rb;
> +               while (rb) {
> +                       parent = rb_entry(rb, struct drm_mm_node, rb);
> +                       if (parent->__subtree_last >= node->__subtree_last)
> +                               break;
> +
> +                       parent->__subtree_last = node->__subtree_last;
> +                       rb = rb_parent(rb);
> +               }
> +
> +               rb = &hole_node->rb;
> +               link = &hole_node->rb.rb_right;
> +       } else {
> +               rb = NULL;
> +               link = &mm->interval_tree.rb_node;
> +       }
> +
> +       while (*link) {
> +               rb = *link;
> +               parent = rb_entry(rb, struct drm_mm_node, rb);
> +               if (parent->__subtree_last < node->__subtree_last)
> +                       parent->__subtree_last = node->__subtree_last;
> +               if (node->start < parent->start)
> +                       link = &parent->rb.rb_left;
> +               else
> +                       link = &parent->rb.rb_right;
> +       }
> +
> +       rb_link_node(&node->rb, rb, link);
> +       rb_insert_augmented(&node->rb,
> +                           &mm->interval_tree,
> +                           &drm_mm_interval_tree_augment);
> +}
> +
>  static void drm_mm_insert_helper(struct drm_mm_node *hole_node,
>                                  struct drm_mm_node *node,
>                                  u64 size, unsigned alignment,
> @@ -153,6 +220,8 @@ static void drm_mm_insert_helper(struct drm_mm_node *hole_node,
>         INIT_LIST_HEAD(&node->hole_stack);
>         list_add(&node->node_list, &hole_node->node_list);
>
> +       drm_mm_interval_tree_add_node(hole_node, node);
> +
>         BUG_ON(node->start + node->size > adj_end);
>
>         node->hole_follows = 0;
> @@ -178,41 +247,52 @@ static void drm_mm_insert_helper(struct drm_mm_node *hole_node,
>   */
>  int drm_mm_reserve_node(struct drm_mm *mm, struct drm_mm_node *node)
>  {
> +       u64 end = node->start + node->size;
>         struct drm_mm_node *hole;
> -       u64 end;
> -       u64 hole_start;
> -       u64 hole_end;
> -
> -       BUG_ON(node == NULL);
> +       u64 hole_start, hole_end;
>
>         end = node->start + node->size;
>
>         /* Find the relevant hole to add our node to */
> -       drm_mm_for_each_hole(hole, mm, hole_start, hole_end) {
> -               if (hole_start > node->start || hole_end < end)
> -                       continue;
> +       hole = drm_mm_interval_tree_iter_first(&mm->interval_tree,
> +                                              node->start, ~(u64)0);
> +       if (hole) {
> +               if (hole->start < end)
> +                       return -ENOSPC;
> +       } else {
> +               hole = list_entry(&mm->head_node.node_list,
> +                                 typeof(*hole), node_list);
> +       }
>
> -               node->mm = mm;
> -               node->allocated = 1;
> +       hole = list_last_entry(&hole->node_list, typeof(*hole), node_list);
> +       if (!hole->hole_follows)
> +               return -ENOSPC;
>
> -               INIT_LIST_HEAD(&node->hole_stack);
> -               list_add(&node->node_list, &hole->node_list);
> +       hole_start = __drm_mm_hole_node_start(hole);
> +       hole_end = __drm_mm_hole_node_end(hole);
> +       if (hole_start > node->start || hole_end < end)
> +               return -ENOSPC;
>
> -               if (node->start == hole_start) {
> -                       hole->hole_follows = 0;
> -                       list_del_init(&hole->hole_stack);
> -               }
> +       node->mm = mm;
> +       node->allocated = 1;
>
> -               node->hole_follows = 0;
> -               if (end != hole_end) {
> -                       list_add(&node->hole_stack, &mm->hole_stack);
> -                       node->hole_follows = 1;
> -               }
> +       INIT_LIST_HEAD(&node->hole_stack);
> +       list_add(&node->node_list, &hole->node_list);
>
> -               return 0;
> +       drm_mm_interval_tree_add_node(hole, node);
> +
> +       if (node->start == hole_start) {
> +               hole->hole_follows = 0;
> +               list_del_init(&hole->hole_stack);
> +       }
> +
> +       node->hole_follows = 0;
> +       if (end != hole_end) {
> +               list_add(&node->hole_stack, &mm->hole_stack);
> +               node->hole_follows = 1;
>         }
>
> -       return -ENOSPC;
> +       return 0;
>  }
>  EXPORT_SYMBOL(drm_mm_reserve_node);
>
> @@ -302,6 +382,8 @@ static void drm_mm_insert_helper_range(struct drm_mm_node *hole_node,
>         INIT_LIST_HEAD(&node->hole_stack);
>         list_add(&node->node_list, &hole_node->node_list);
>
> +       drm_mm_interval_tree_add_node(hole_node, node);
> +
>         BUG_ON(node->start < start);
>         BUG_ON(node->start < adj_start);
>         BUG_ON(node->start + node->size > adj_end);
> @@ -390,6 +472,7 @@ void drm_mm_remove_node(struct drm_mm_node *node)
>         } else
>                 list_move(&prev_node->hole_stack, &mm->hole_stack);
>
> +       drm_mm_interval_tree_remove(node, &mm->interval_tree);
>         list_del(&node->node_list);
>         node->allocated = 0;
>  }
> @@ -516,11 +599,13 @@ void drm_mm_replace_node(struct drm_mm_node *old, struct drm_mm_node *new)
>  {
>         list_replace(&old->node_list, &new->node_list);
>         list_replace(&old->hole_stack, &new->hole_stack);
> +       rb_replace_node(&old->rb, &new->rb, &old->mm->interval_tree);
>         new->hole_follows = old->hole_follows;
>         new->mm = old->mm;
>         new->start = old->start;
>         new->size = old->size;
>         new->color = old->color;
> +       new->__subtree_last = old->__subtree_last;
>
>         old->allocated = 0;
>         new->allocated = 1;
> @@ -758,6 +843,8 @@ void drm_mm_init(struct drm_mm * mm, u64 start, u64 size)
>         mm->head_node.size = start - mm->head_node.start;
>         list_add_tail(&mm->head_node.hole_stack, &mm->hole_stack);
>
> +       mm->interval_tree = RB_ROOT;
> +
>         mm->color_adjust = NULL;
>  }
>  EXPORT_SYMBOL(drm_mm_init);
> diff --git a/include/drm/drm_mm.h b/include/drm/drm_mm.h
> index fc65118e5077..205ddcf6d55d 100644
> --- a/include/drm/drm_mm.h
> +++ b/include/drm/drm_mm.h
> @@ -37,6 +37,7 @@
>   * Generic range manager structs
>   */
>  #include <linux/bug.h>
> +#include <linux/rbtree.h>
>  #include <linux/kernel.h>
>  #include <linux/list.h>
>  #include <linux/spinlock.h>
> @@ -61,6 +62,7 @@ enum drm_mm_allocator_flags {
>  struct drm_mm_node {
>         struct list_head node_list;
>         struct list_head hole_stack;
> +       struct rb_node rb;
>         unsigned hole_follows : 1;
>         unsigned scanned_block : 1;
>         unsigned scanned_prev_free : 1;
> @@ -70,6 +72,7 @@ struct drm_mm_node {
>         unsigned long color;
>         u64 start;
>         u64 size;
> +       u64 __subtree_last;
>         struct drm_mm *mm;
>  };
>
> @@ -79,6 +82,9 @@ struct drm_mm {
>         /* head_node.node_list is the list of all memory nodes, ordered
>          * according to the (increasing) start address of the memory node. */
>         struct drm_mm_node head_node;
> +       /* Keep an interval_tree for fast lookup of drm_mm_nodes by address. */
> +       struct rb_root interval_tree;
> +
>         unsigned int scan_check_range : 1;
>         unsigned scan_alignment;
>         unsigned long scan_color;
> @@ -295,6 +301,12 @@ void drm_mm_init(struct drm_mm *mm,
>  void drm_mm_takedown(struct drm_mm *mm);
>  bool drm_mm_clean(struct drm_mm *mm);
>
> +struct drm_mm_node *
> +drm_mm_interval_first(struct drm_mm *mm, u64 start, u64 last);
> +
> +struct drm_mm_node *
> +drm_mm_interval_next(struct drm_mm_node *node, u64 start, u64 last);
> +
>  void drm_mm_init_scan(struct drm_mm *mm,
>                       u64 size,
>                       unsigned alignment,
> --
> 2.8.1
>


More information about the Intel-gfx mailing list