[Intel-gfx] [PATCH 1/6] drm/i915/ttm: add ttm_buddy_man

Matthew Auld matthew.auld at intel.com
Tue Jun 8 08:11:05 UTC 2021


On 08/06/2021 08:11, Thomas Hellström wrote:
> On Mon, 2021-06-07 at 19:22 +0100, Matthew Auld wrote:
>> Add back our standalone i915_buddy allocator and integrate it into a
>> ttm_resource_manager. This will plug into our ttm backend for
>> managing
>> device local-memory in the next couple of patches.
>>
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Thomas Hellström <thomas.hellstrom at linux.intel.com>
>> ---
>>
> 
> Since the buddy + selftests have been part of the driver before, I
> didn't review them separately, but for the TTM interface, some minor
> comments below. With those fixed,
> 
> Acked-by: Thomas Hellström <thomas.hellstrom at linux.intel.com>
> 
> 
>> diff --git a/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
>> b/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
>> new file mode 100644
>> index 000000000000..d7bf37be1932
>> --- /dev/null
>> +++ b/drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
>> @@ -0,0 +1,246 @@
>> +// SPDX-License-Identifier: MIT
>> +/*
>> + * Copyright © 2021 Intel Corporation
>> + */
>> +
>> +#include <linux/slab.h>
>> +
>> +#include <drm/ttm/ttm_bo_driver.h>
>> +#include <drm/ttm/ttm_placement.h>
>> +
>> +#include "i915_ttm_buddy_manager.h"
>> +
>> +#include "i915_buddy.h"
>> +#include "i915_gem.h"
>> +
>> +struct i915_ttm_buddy_manager {
>> +       struct ttm_resource_manager manager;
>> +       struct i915_buddy_mm mm;
>> +       struct list_head reserved;
>> +       struct mutex lock;
>> +};
>> +
>> +static inline struct i915_ttm_buddy_manager *
> 
> "inline" shouldn't be needed here.
> 
>> +to_buddy_manager(struct ttm_resource_manager *man)
>> +{
>> +       return container_of(man, struct i915_ttm_buddy_manager,
>> manager);
>> +}
>> +
>> +static int i915_ttm_buddy_man_alloc(struct ttm_resource_manager
>> *man,
>> +                                   struct ttm_buffer_object *bo,
>> +                                   const struct ttm_place *place,
>> +                                   struct ttm_resource **res)
>> +{
>> +       struct i915_ttm_buddy_manager *bman = to_buddy_manager(man);
>> +       struct i915_ttm_buddy_resource *bman_res;
>> +       struct i915_buddy_mm *mm = &bman->mm;
>> +       unsigned long n_pages;
>> +       unsigned int min_order;
>> +       u64 size;
>> +       int err;
>> +
>> +       GEM_BUG_ON(place->fpfn || place->lpfn);
>> +       GEM_BUG_ON(bo->page_alignment < mm->chunk_size);
>> +
>> +       bman_res = kzalloc(sizeof(*bman_res), GFP_KERNEL);
>> +       if (!bman_res)
>> +               return -ENOMEM;
>> +
>> +       ttm_resource_init(bo, place, &bman_res->base);
>> +       INIT_LIST_HEAD(&bman_res->blocks);
>> +       bman_res->mm = mm;
>> +
>> +       GEM_BUG_ON(!bman_res->base.num_pages);
>> +       size = bman_res->base.num_pages << PAGE_SHIFT;
>> +
>> +       min_order = ilog2(bo->page_alignment) - ilog2(mm-
>>> chunk_size);
>> +       if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
>> +               size = roundup_pow_of_two(size);
>> +               min_order = ilog2(size) - ilog2(mm->chunk_size);
>> +       }
>> +
>> +       if (size > mm->size) {
>> +               err = -E2BIG;
>> +               goto err_free_res;
>> +       }
>> +
>> +       n_pages = size >> ilog2(mm->chunk_size);
>> +
>> +       do {
>> +               struct i915_buddy_block *block;
>> +               unsigned int order;
>> +
>> +               order = fls(n_pages) - 1;
>> +               GEM_BUG_ON(order > mm->max_order);
>> +               GEM_BUG_ON(order < min_order);
>> +
>> +               do {
>> +                       mutex_lock(&bman->lock);
>> +                       block = i915_buddy_alloc(mm, order);
>> +                       mutex_unlock(&bman->lock);
>> +                       if (!IS_ERR(block))
>> +                               break;
>> +
>> +                       if (order-- == min_order) {
>> +                               err = -ENXIO;
> 
> IIRC, TTM relies on -ENOSPC to retry with evictions.

Ah, right. We convert that back to -ENXIO in the upper levels somewhere?

> 
>> +                               goto err_free_blocks;
>> +                       }
>> +               } while (1);
>> +
>> +               n_pages -= BIT(order);
>> +
>> +               list_add_tail(&block->link, &bman_res->blocks);
>> +
>> +               if (!n_pages)
>> +                       break;
>> +       } while (1);
>> +
>> +       *res = &bman_res->base;
>> +       return 0;
>> +
>> +err_free_blocks:
>> +       mutex_lock(&bman->lock);
>> +       i915_buddy_free_list(mm, &bman_res->blocks);
>> +       mutex_unlock(&bman->lock);
>> +err_free_res:
>> +       kfree(bman_res);
>> +       return err;
>> +}
>> +
> 
> /Thomas
> 
> 


More information about the Intel-gfx mailing list