[PATCH] drm/amdgpu: further lower VRAM allocation overhead

Felix Kuehling felix.kuehling at amd.com
Tue Jul 13 16:11:23 UTC 2021


Am 2021-07-13 um 9:32 a.m. schrieb Christian König:
> For allocations larger than 48MiB we need more than a page for the
> housekeeping in the worst case resulting in the usual vmalloc overhead.
>
> Try to avoid this by assuming the good case and only falling back to the
> worst case if this didn't worked.
>
> Signed-off-by: Christian König <christian.koenig at amd.com>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c | 80 +++++++++++++++-----
>  1 file changed, 60 insertions(+), 20 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
> index 2fd77c36a1ff..ab8c5e28df7b 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
> @@ -361,19 +361,23 @@ static void amdgpu_vram_mgr_virt_start(struct ttm_resource *mem,
>   * @man: TTM memory type manager
>   * @tbo: TTM BO we need this range for
>   * @place: placement flags and restrictions
> - * @mem: the resulting mem object
> + * @num_nodes: number of page nodes to use.
> + * @pages_per_node: number of pages per node to use.
> + * @res: the resulting mem object
>   *
>   * Allocate VRAM for the given BO.
>   */
>  static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
>  			       struct ttm_buffer_object *tbo,
>  			       const struct ttm_place *place,
> +			       unsigned long num_nodes,
> +			       unsigned long pages_per_node,
>  			       struct ttm_resource **res)
>  {
> -	unsigned long lpfn, num_nodes, pages_per_node, pages_left, pages;
>  	struct amdgpu_vram_mgr *mgr = to_vram_mgr(man);
>  	struct amdgpu_device *adev = to_amdgpu_device(mgr);
>  	uint64_t vis_usage = 0, mem_bytes, max_bytes;
> +	unsigned long lpfn, pages_left, pages;
>  	struct ttm_range_mgr_node *node;
>  	struct drm_mm *mm = &mgr->mm;
>  	enum drm_mm_insert_mode mode;
> @@ -395,21 +399,6 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
>  		goto error_sub;
>  	}
>  
> -	if (place->flags & TTM_PL_FLAG_CONTIGUOUS) {
> -		pages_per_node = ~0ul;
> -		num_nodes = 1;
> -	} else {
> -#ifdef CONFIG_TRANSPARENT_HUGEPAGE
> -		pages_per_node = HPAGE_PMD_NR;
> -#else
> -		/* default to 2MB */
> -		pages_per_node = 2UL << (20UL - PAGE_SHIFT);
> -#endif
> -		pages_per_node = max_t(uint32_t, pages_per_node,
> -				       tbo->page_alignment);
> -		num_nodes = DIV_ROUND_UP_ULL(PFN_UP(mem_bytes), pages_per_node);
> -	}
> -
>  	node = kvmalloc(struct_size(node, mm_nodes, num_nodes),
>  			GFP_KERNEL | __GFP_ZERO);
>  	if (!node) {
> @@ -431,10 +420,15 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
>  	i = 0;
>  	spin_lock(&mgr->lock);
>  	while (pages_left) {
> -		uint32_t alignment = tbo->page_alignment;
> +		unsigned long alignment = tbo->page_alignment;
> +
> +		if (i >= num_nodes) {
> +			r = -E2BIG;
> +			goto error_free;
> +		}
>  
>  		if (pages >= pages_per_node)
> -			alignment = pages_per_node;
> +			alignment = max(alignment, pages_per_node);

I don't understand this change. Is this an unrelated fix? pages_per_node
is already bumped up to tbo->page_alignment in amdgpu_vram_mgr_alloc. So
this "max" operation here seems redundant.

Other than that, the patch is

Reviewed-by: Felix Kuehling <Felix.Kuehling at amd.com>

@JinHuiEric, can you confirm the performance improvement?

Thanks,
  Felix


>  
>  		r = drm_mm_insert_node_in_range(mm, &node->mm_nodes[i], pages,
>  						alignment, 0, place->fpfn,
> @@ -483,6 +477,52 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
>  	return r;
>  }
>  
> +/**
> + * amdgpu_vram_mgr_alloc - allocate new range
> + *
> + * @man: TTM memory type manager
> + * @tbo: TTM BO we need this range for
> + * @place: placement flags and restrictions
> + * @res: the resulting mem object
> + *
> + * Allocate VRAM for the given BO.
> + */
> +static int amdgpu_vram_mgr_alloc(struct ttm_resource_manager *man,
> +				 struct ttm_buffer_object *tbo,
> +				 const struct ttm_place *place,
> +				 struct ttm_resource **res)
> +{
> +	unsigned long num_nodes, pages_per_node;
> +	struct ttm_range_mgr_node *node;
> +	int r;
> +
> +	if (place->flags & TTM_PL_FLAG_CONTIGUOUS)
> +		return amdgpu_vram_mgr_new(man, tbo, place, 1, ~0ul, res);
> +
> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE
> +	pages_per_node = HPAGE_PMD_NR;
> +#else
> +	/* default to 2MB */
> +	pages_per_node = 2UL << (20UL - PAGE_SHIFT);
> +#endif
> +	pages_per_node = max_t(uint32_t, pages_per_node, tbo->page_alignment);
> +	num_nodes = DIV_ROUND_UP_ULL(PFN_UP(tbo->base.size), pages_per_node);
> +
> +	if (struct_size(node, mm_nodes, num_nodes) > PAGE_SIZE) {
> +		size_t size = PAGE_SIZE;
> +
> +		size -= sizeof(struct ttm_range_mgr_node);
> +		size /= sizeof(struct drm_mm_node);
> +		r = amdgpu_vram_mgr_new(man, tbo, place, size, pages_per_node,
> +					res);
> +		if (r != -E2BIG)
> +			return r;
> +	}
> +
> +	return amdgpu_vram_mgr_new(man, tbo, place, num_nodes, pages_per_node,
> +				   res);
> +}
> +
>  /**
>   * amdgpu_vram_mgr_del - free ranges
>   *
> @@ -680,7 +720,7 @@ static void amdgpu_vram_mgr_debug(struct ttm_resource_manager *man,
>  }
>  
>  static const struct ttm_resource_manager_func amdgpu_vram_mgr_func = {
> -	.alloc	= amdgpu_vram_mgr_new,
> +	.alloc	= amdgpu_vram_mgr_alloc,
>  	.free	= amdgpu_vram_mgr_del,
>  	.debug	= amdgpu_vram_mgr_debug
>  };


More information about the amd-gfx mailing list