[PATCH 1/6] drm/i915/ttm: add ttm_buddy_man
Thomas Hellström
thomas.hellstrom at linux.intel.com
Tue Jun 8 07:11:37 UTC 2021
On Mon, 2021-06-07 at 19:22 +0100, Matthew Auld wrote:
> Add back our standalone i915_buddy allocator and integrate it into a
> ttm_resource_manager. This will plug into our ttm backend for
> managing
> device local-memory in the next couple of patches.
>
> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> Cc: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> ---
>
Since the buddy + selftests have been part of the driver before, I
didn't review them separately, but for the TTM interface, some minor
comments below. With those fixed,
Acked-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> diff --git a/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
> b/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
> new file mode 100644
> index 000000000000..d7bf37be1932
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
> @@ -0,0 +1,246 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +#include <linux/slab.h>
> +
> +#include <drm/ttm/ttm_bo_driver.h>
> +#include <drm/ttm/ttm_placement.h>
> +
> +#include "i915_ttm_buddy_manager.h"
> +
> +#include "i915_buddy.h"
> +#include "i915_gem.h"
> +
> +struct i915_ttm_buddy_manager {
> + struct ttm_resource_manager manager;
> + struct i915_buddy_mm mm;
> + struct list_head reserved;
> + struct mutex lock;
> +};
> +
> +static inline struct i915_ttm_buddy_manager *
"inline" shouldn't be needed here.
> +to_buddy_manager(struct ttm_resource_manager *man)
> +{
> + return container_of(man, struct i915_ttm_buddy_manager,
> manager);
> +}
> +
> +static int i915_ttm_buddy_man_alloc(struct ttm_resource_manager
> *man,
> + struct ttm_buffer_object *bo,
> + const struct ttm_place *place,
> + struct ttm_resource **res)
> +{
> + struct i915_ttm_buddy_manager *bman = to_buddy_manager(man);
> + struct i915_ttm_buddy_resource *bman_res;
> + struct i915_buddy_mm *mm = &bman->mm;
> + unsigned long n_pages;
> + unsigned int min_order;
> + u64 size;
> + int err;
> +
> + GEM_BUG_ON(place->fpfn || place->lpfn);
> + GEM_BUG_ON(bo->page_alignment < mm->chunk_size);
> +
> + bman_res = kzalloc(sizeof(*bman_res), GFP_KERNEL);
> + if (!bman_res)
> + return -ENOMEM;
> +
> + ttm_resource_init(bo, place, &bman_res->base);
> + INIT_LIST_HEAD(&bman_res->blocks);
> + bman_res->mm = mm;
> +
> + GEM_BUG_ON(!bman_res->base.num_pages);
> + size = bman_res->base.num_pages << PAGE_SHIFT;
> +
> + min_order = ilog2(bo->page_alignment) - ilog2(mm-
> >chunk_size);
> + if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
> + size = roundup_pow_of_two(size);
> + min_order = ilog2(size) - ilog2(mm->chunk_size);
> + }
> +
> + if (size > mm->size) {
> + err = -E2BIG;
> + goto err_free_res;
> + }
> +
> + n_pages = size >> ilog2(mm->chunk_size);
> +
> + do {
> + struct i915_buddy_block *block;
> + unsigned int order;
> +
> + order = fls(n_pages) - 1;
> + GEM_BUG_ON(order > mm->max_order);
> + GEM_BUG_ON(order < min_order);
> +
> + do {
> + mutex_lock(&bman->lock);
> + block = i915_buddy_alloc(mm, order);
> + mutex_unlock(&bman->lock);
> + if (!IS_ERR(block))
> + break;
> +
> + if (order-- == min_order) {
> + err = -ENXIO;
IIRC, TTM relies on -ENOSPC to retry with evictions.
> + goto err_free_blocks;
> + }
> + } while (1);
> +
> + n_pages -= BIT(order);
> +
> + list_add_tail(&block->link, &bman_res->blocks);
> +
> + if (!n_pages)
> + break;
> + } while (1);
> +
> + *res = &bman_res->base;
> + return 0;
> +
> +err_free_blocks:
> + mutex_lock(&bman->lock);
> + i915_buddy_free_list(mm, &bman_res->blocks);
> + mutex_unlock(&bman->lock);
> +err_free_res:
> + kfree(bman_res);
> + return err;
> +}
> +
/Thomas
More information about the dri-devel
mailing list