[PATCH 1/2] drm/etnaviv: Preallocate STLB according to CPU PAGE_SIZE
Lucas Stach
l.stach at pengutronix.de
Tue Dec 3 18:04:18 UTC 2024
Am Freitag, dem 08.11.2024 um 22:36 +0800 schrieb Sui Jingfeng:
> The dma_direct_alloc() allocate one page at minmium, which size is the CPU
> PAGE_SIZE. while the etnaviv_iommuv2_ensure_stlb() only ask for 4KiB. The
> rest memory space that beyond 4KiB gets wasted on bigger page size systems.
> For example, on 16KiB CPU page size systems, we will waste the rest 12KiB.
> On 64KiB CPU page size systems, we will waste the rest 60KiB.
>
> Since addresses within one page are always contiguous, the rest memory can
> be used to store adjacent slave TLB entries. Then, when the neighbourhoods
> TLB is being hit on the next time, we don't have to ask another one page
> from the system. Saving both memorys and times overhead because of that.
>
While this isn't adding a lot of code to etnaviv, I wonder if this
couldn't be handled by using a dma_pool for the pagetable allocations.
Regards,
Lucas
> Signed-off-by: Sui Jingfeng <sui.jingfeng at linux.dev>
> ---
> drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c | 64 +++++++++++++++++++---
> 1 file changed, 56 insertions(+), 8 deletions(-)
>
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c b/drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
> index d664ae29ae20..fa6eed1ae1be 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
> @@ -44,19 +44,66 @@ to_v2_context(struct etnaviv_iommu_context *context)
> return container_of(context, struct etnaviv_iommuv2_context, base);
> }
>
> +static int etnaviv_iommuv2_stlb_free(struct etnaviv_iommuv2_context *context)
> +{
> + struct device *dev = context->base.global->dev;
> + unsigned int i;
> +
> + for (i = 0; i < MMUv2_MAX_STLB_ENTRIES; ++i) {
> + u32 *vaddr = context->stlb_cpu[i];
> +
> + if (!vaddr)
> + continue;
> +
> + context->stlb_cpu[i] = NULL;
> +
> + if (i % (PAGE_SIZE / SZ_4K))
> + continue;
> +
> + dma_free_wc(dev, PAGE_SIZE, vaddr, context->stlb_dma[i]);
> + }
> +
> + return 0;
> +}
> +
> +static int
> +etnaviv_iommuv2_ensure_stlb_new(struct etnaviv_iommuv2_context *context,
> + unsigned int stlb)
> +{
> + struct device *dev = context->base.global->dev;
> + void *vaddr;
> + dma_addr_t daddr;
> + unsigned int i;
> +
> + if (context->stlb_cpu[stlb])
> + return 0;
> +
> + vaddr = dma_alloc_wc(dev, PAGE_SIZE, &daddr, GFP_KERNEL);
> + if (!vaddr)
> + return -ENOMEM;
> +
> + memset32(vaddr, MMUv2_PTE_EXCEPTION, PAGE_SIZE / sizeof(u32));
> +
> + stlb &= ~(PAGE_SIZE / SZ_4K - 1);
> +
> + for (i = 0; i < PAGE_SIZE / SZ_4K; ++i) {
> + context->stlb_cpu[stlb + i] = vaddr;
> + context->stlb_dma[stlb + i] = daddr;
> + context->mtlb_cpu[stlb + i] = daddr | MMUv2_PTE_PRESENT;
> + vaddr += SZ_4K;
> + daddr += SZ_4K;
> + }
> +
> + return 0;
> +}
> +
> static void etnaviv_iommuv2_free(struct etnaviv_iommu_context *context)
> {
> struct etnaviv_iommuv2_context *v2_context = to_v2_context(context);
> - int i;
>
> drm_mm_takedown(&context->mm);
>
> - for (i = 0; i < MMUv2_MAX_STLB_ENTRIES; i++) {
> - if (v2_context->stlb_cpu[i])
> - dma_free_wc(context->global->dev, SZ_4K,
> - v2_context->stlb_cpu[i],
> - v2_context->stlb_dma[i]);
> - }
> + etnaviv_iommuv2_stlb_free(v2_context);
>
> dma_free_wc(context->global->dev, SZ_4K, v2_context->mtlb_cpu,
> v2_context->mtlb_dma);
> @@ -65,6 +112,7 @@ static void etnaviv_iommuv2_free(struct etnaviv_iommu_context *context)
>
> vfree(v2_context);
> }
> +
> static int
> etnaviv_iommuv2_ensure_stlb(struct etnaviv_iommuv2_context *v2_context,
> int stlb)
> @@ -109,7 +157,7 @@ static int etnaviv_iommuv2_map(struct etnaviv_iommu_context *context,
> mtlb_entry = (iova & MMUv2_MTLB_MASK) >> MMUv2_MTLB_SHIFT;
> stlb_entry = (iova & MMUv2_STLB_MASK) >> MMUv2_STLB_SHIFT;
>
> - ret = etnaviv_iommuv2_ensure_stlb(v2_context, mtlb_entry);
> + ret = etnaviv_iommuv2_ensure_stlb_new(v2_context, mtlb_entry);
> if (ret)
> return ret;
>
More information about the dri-devel
mailing list