[PATCH] drm/xe/vram: drop 2G block restriction
Matthew Auld
matthew.auld at intel.com
Fri Nov 15 08:53:58 UTC 2024
On 14/11/2024 14:22, Matthew Brost wrote:
> On Wed, Nov 13, 2024 at 05:23:47PM +0000, Matthew Auld wrote:
>> Currently we limit the max block size for all users to ensure each block
>> can fit within a sg entry (uint). Drop this restriction and tweak the sg
>> construction to instead handle this itself and break down blocks which
>> are too big, if needed. Most users don't need an sg list in the first
>> place.
>>
>
> Code looks correct. Curious what the motivation for the series as before
> / after the series everything is functional. Just cleaning up a FIXME?
The motivation came from vlk-64377. But regardless of that I was
thinking that addressing the FIXME here was not a bad idea.
>
> Anyways LGTM:
> Reviewed-by: Matthew Brost <matthew.brost at intel.com>
Thanks.
>
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Satyanarayana K V P <satyanarayana.k.v.p at intel.com>
>> Cc: Matthew Brost <matthew.brost at intel.com>
>> ---
>> drivers/gpu/drm/xe/xe_ttm_vram_mgr.c | 51 +++++++---------------------
>> 1 file changed, 12 insertions(+), 39 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c b/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> index 423b261ea743..1d39a8c53b3a 100644
>> --- a/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> +++ b/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> @@ -52,7 +52,7 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>> struct xe_ttm_vram_mgr *mgr = to_xe_ttm_vram_mgr(man);
>> struct xe_ttm_vram_mgr_resource *vres;
>> struct drm_buddy *mm = &mgr->mm;
>> - u64 size, remaining_size, min_page_size;
>> + u64 size, min_page_size;
>> unsigned long lpfn;
>> int err;
>>
>> @@ -98,17 +98,6 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>> goto error_fini;
>> }
>>
>> - if (WARN_ON(min_page_size > SZ_2G)) { /* FIXME: sg limit */
>> - err = -EINVAL;
>> - goto error_fini;
>> - }
>> -
>> - if (WARN_ON((size > SZ_2G &&
>> - (vres->base.placement & TTM_PL_FLAG_CONTIGUOUS)))) {
>> - err = -EINVAL;
>> - goto error_fini;
>> - }
>> -
>> if (WARN_ON(!IS_ALIGNED(size, min_page_size))) {
>> err = -EINVAL;
>> goto error_fini;
>> @@ -116,9 +105,8 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>
>> mutex_lock(&mgr->lock);
>> if (lpfn <= mgr->visible_size >> PAGE_SHIFT && size > mgr->visible_avail) {
>> - mutex_unlock(&mgr->lock);
>> err = -ENOSPC;
>> - goto error_fini;
>> + goto error_unlock;
>> }
>>
>> if (place->fpfn + (size >> PAGE_SHIFT) != place->lpfn &&
>> @@ -129,25 +117,11 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>> lpfn = max_t(unsigned long, place->fpfn + (size >> PAGE_SHIFT), lpfn);
>> }
>>
>> - remaining_size = size;
>> - do {
>> - /*
>> - * Limit maximum size to 2GiB due to SG table limitations.
>> - * FIXME: Should maybe be handled as part of sg construction.
>> - */
>> - u64 alloc_size = min_t(u64, remaining_size, SZ_2G);
>> -
>> - err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
>> - (u64)lpfn << PAGE_SHIFT,
>> - alloc_size,
>> - min_page_size,
>> - &vres->blocks,
>> - vres->flags);
>> - if (err)
>> - goto error_free_blocks;
>> -
>> - remaining_size -= alloc_size;
>> - } while (remaining_size);
>> + err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
>> + (u64)lpfn << PAGE_SHIFT, size,
>> + min_page_size, &vres->blocks, vres->flags);
>> + if (err)
>> + goto error_unlock;
>>
>> if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
>> if (!drm_buddy_block_trim(mm, NULL, vres->base.size, &vres->blocks))
>> @@ -194,9 +168,7 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>
>> *res = &vres->base;
>> return 0;
>> -
>> -error_free_blocks:
>> - drm_buddy_free_list(mm, &vres->blocks, 0);
>> +error_unlock:
>> mutex_unlock(&mgr->lock);
>> error_fini:
>> ttm_resource_fini(man, &vres->base);
>> @@ -393,7 +365,8 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>> xe_res_first(res, offset, length, &cursor);
>> while (cursor.remaining) {
>> num_entries++;
>> - xe_res_next(&cursor, cursor.size);
>> + /* Limit maximum size to 2GiB due to SG table limitations. */
>> + xe_res_next(&cursor, min_t(u64, cursor.size, SZ_2G));
>> }
>>
>> r = sg_alloc_table(*sgt, num_entries, GFP_KERNEL);
>> @@ -413,7 +386,7 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>> xe_res_first(res, offset, length, &cursor);
>> for_each_sgtable_sg((*sgt), sg, i) {
>> phys_addr_t phys = cursor.start + tile->mem.vram.io_start;
>> - size_t size = cursor.size;
>> + size_t size = min_t(u64, cursor.size, SZ_2G);
>> dma_addr_t addr;
>>
>> addr = dma_map_resource(dev, phys, size, dir,
>> @@ -426,7 +399,7 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>> sg_dma_address(sg) = addr;
>> sg_dma_len(sg) = size;
>>
>> - xe_res_next(&cursor, cursor.size);
>> + xe_res_next(&cursor, size);
>> }
>>
>> return 0;
>> --
>> 2.47.0
>>
More information about the Intel-xe
mailing list