[PATCH] drm/xe/vram: drop 2G block restriction

Matthew Auld matthew.auld at intel.com
Fri Nov 15 08:53:58 UTC 2024


On 14/11/2024 14:22, Matthew Brost wrote:
> On Wed, Nov 13, 2024 at 05:23:47PM +0000, Matthew Auld wrote:
>> Currently we limit the max block size for all users to ensure each block
>> can fit within a sg entry (uint). Drop this restriction and tweak the sg
>> construction to instead handle this itself and break down blocks which
>> are too big, if needed. Most users don't need an sg list in the first
>> place.
>>
> 
> Code looks correct. Curious what the motivation for the series as before
> / after the series everything is functional. Just cleaning up a FIXME?

The motivation came from vlk-64377. But regardless of that I was 
thinking that addressing the FIXME here was not a bad idea.

> 
> Anyways LGTM:
> Reviewed-by: Matthew Brost <matthew.brost at intel.com>

Thanks.

> 
>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
>> Cc: Satyanarayana K V P <satyanarayana.k.v.p at intel.com>
>> Cc: Matthew Brost <matthew.brost at intel.com>
>> ---
>>   drivers/gpu/drm/xe/xe_ttm_vram_mgr.c | 51 +++++++---------------------
>>   1 file changed, 12 insertions(+), 39 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c b/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> index 423b261ea743..1d39a8c53b3a 100644
>> --- a/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> +++ b/drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
>> @@ -52,7 +52,7 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>   	struct xe_ttm_vram_mgr *mgr = to_xe_ttm_vram_mgr(man);
>>   	struct xe_ttm_vram_mgr_resource *vres;
>>   	struct drm_buddy *mm = &mgr->mm;
>> -	u64 size, remaining_size, min_page_size;
>> +	u64 size, min_page_size;
>>   	unsigned long lpfn;
>>   	int err;
>>   
>> @@ -98,17 +98,6 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>   		goto error_fini;
>>   	}
>>   
>> -	if (WARN_ON(min_page_size > SZ_2G)) { /* FIXME: sg limit */
>> -		err = -EINVAL;
>> -		goto error_fini;
>> -	}
>> -
>> -	if (WARN_ON((size > SZ_2G &&
>> -		     (vres->base.placement & TTM_PL_FLAG_CONTIGUOUS)))) {
>> -		err = -EINVAL;
>> -		goto error_fini;
>> -	}
>> -
>>   	if (WARN_ON(!IS_ALIGNED(size, min_page_size))) {
>>   		err = -EINVAL;
>>   		goto error_fini;
>> @@ -116,9 +105,8 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>   
>>   	mutex_lock(&mgr->lock);
>>   	if (lpfn <= mgr->visible_size >> PAGE_SHIFT && size > mgr->visible_avail) {
>> -		mutex_unlock(&mgr->lock);
>>   		err = -ENOSPC;
>> -		goto error_fini;
>> +		goto error_unlock;
>>   	}
>>   
>>   	if (place->fpfn + (size >> PAGE_SHIFT) != place->lpfn &&
>> @@ -129,25 +117,11 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>   		lpfn = max_t(unsigned long, place->fpfn + (size >> PAGE_SHIFT), lpfn);
>>   	}
>>   
>> -	remaining_size = size;
>> -	do {
>> -		/*
>> -		 * Limit maximum size to 2GiB due to SG table limitations.
>> -		 * FIXME: Should maybe be handled as part of sg construction.
>> -		 */
>> -		u64 alloc_size = min_t(u64, remaining_size, SZ_2G);
>> -
>> -		err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
>> -					     (u64)lpfn << PAGE_SHIFT,
>> -					     alloc_size,
>> -					     min_page_size,
>> -					     &vres->blocks,
>> -					     vres->flags);
>> -		if (err)
>> -			goto error_free_blocks;
>> -
>> -		remaining_size -= alloc_size;
>> -	} while (remaining_size);
>> +	err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
>> +				     (u64)lpfn << PAGE_SHIFT, size,
>> +				     min_page_size, &vres->blocks, vres->flags);
>> +	if (err)
>> +		goto error_unlock;
>>   
>>   	if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
>>   		if (!drm_buddy_block_trim(mm, NULL, vres->base.size, &vres->blocks))
>> @@ -194,9 +168,7 @@ static int xe_ttm_vram_mgr_new(struct ttm_resource_manager *man,
>>   
>>   	*res = &vres->base;
>>   	return 0;
>> -
>> -error_free_blocks:
>> -	drm_buddy_free_list(mm, &vres->blocks, 0);
>> +error_unlock:
>>   	mutex_unlock(&mgr->lock);
>>   error_fini:
>>   	ttm_resource_fini(man, &vres->base);
>> @@ -393,7 +365,8 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>>   	xe_res_first(res, offset, length, &cursor);
>>   	while (cursor.remaining) {
>>   		num_entries++;
>> -		xe_res_next(&cursor, cursor.size);
>> +		/* Limit maximum size to 2GiB due to SG table limitations. */
>> +		xe_res_next(&cursor, min_t(u64, cursor.size, SZ_2G));
>>   	}
>>   
>>   	r = sg_alloc_table(*sgt, num_entries, GFP_KERNEL);
>> @@ -413,7 +386,7 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>>   	xe_res_first(res, offset, length, &cursor);
>>   	for_each_sgtable_sg((*sgt), sg, i) {
>>   		phys_addr_t phys = cursor.start + tile->mem.vram.io_start;
>> -		size_t size = cursor.size;
>> +		size_t size = min_t(u64, cursor.size, SZ_2G);
>>   		dma_addr_t addr;
>>   
>>   		addr = dma_map_resource(dev, phys, size, dir,
>> @@ -426,7 +399,7 @@ int xe_ttm_vram_mgr_alloc_sgt(struct xe_device *xe,
>>   		sg_dma_address(sg) = addr;
>>   		sg_dma_len(sg) = size;
>>   
>> -		xe_res_next(&cursor, cursor.size);
>> +		xe_res_next(&cursor, size);
>>   	}
>>   
>>   	return 0;
>> -- 
>> 2.47.0
>>



More information about the Intel-xe mailing list