[PATCH] drm/etnaviv: Print error message if inserting IOVA address range fails
Lucas Stach
l.stach at pengutronix.de
Wed Oct 2 08:44:38 UTC 2024
Am Mittwoch, dem 02.10.2024 um 03:26 +0800 schrieb Sui Jingfeng:
> Hi,
>
> On 2024/10/1 16:27, Lucas Stach wrote:
> > Hi Sui,
> >
> > Am Dienstag, dem 01.10.2024 um 06:17 +0800 schrieb Sui Jingfeng:
> > > Etnaviv assumes that GPU page size is 4KiB, yet on some systems, the CPU
> > > page size is 16 KiB. The size of etnaviv buffer objects will be aligned
> > > to CPU page size on kernel side, however, userspace still assumes the
> > > page size is 4KiB and doing allocation with 4KiB page as unit. This
> > > results in softpin(userspace managed per-process address spaces) fails.
> > > Because kernel side BO takes up bigger address space than user space
> > > assumes whenever the size of a BO is not CPU page size aligned.
> > >
> > Seems we need to track the GPU and CPU allocation sizes separately.
>
>
> The idea is cool and fancy, I have been tried.
>
> By adding a 'user_size' member into the struct etnaviv_gem_object,
> and use this 'user_size'; to track the actual size that user-space
> thing of. (or in other words, the actual size that potential user
> allow to use)
>
> Using 'user_size' is pin, this partly solve VA address space collision
> under softpin fashion. This is partly works under my hasty test. But ...
>
> > Userspace is correct in assuming that the GPU page size is 4K and
> > buffers are aligned to this granule.
>
>
> Vivante GPU support 4KB and 64KB GPU page size.
>
64k is very impractical, as it can't really be mixed freely with 4k
page size. If we ever going to support this, then it will be exposed
via a userspace queryable param, so userspace will know when we need
bigger alignment.
>
> > There should be no need to waste GPU VA space
>
>
> We have nearly 4GBGPU VA space, As far as I can see it, we only use a few. So, is it true
> that we are wealthy about the VA space?
>
Those GPUs are used within systems that support more physical memory
than that. Especially the machines that are using larger page sizes on
the CPU are likely to support more physical memory than 4GB. I don't
see a reason why we should waste GPU VA space when we can avoid it.
It's true that workloads you would run on a GC1000 are unlikely to use
more than a fraction of the GPU VA space, but I can easily see more
capable GPU cores with compute capabilities running up against the GPU
VA space limit.
>
> > just because the CPU page size is larger than that and we
> > need to overallocate buffers to suit the CPU.
>
>
> A single CPU page share the same caching property, therefore, I image that
> asingle VA address range at least should occupy entire room of a single CPU
> page.
>
> Otherwise, it possible that 4 GPUVA share a single CPU page.
> if each GPUVA mapped with a different caching property from others.
> This get coherency requirements involved.
>
That won't happen. We still allocate the whole 16k page for a BO, so no
other BO with different caching flags can share the same CPU page. We
just don't map the whole page to the GPU side.
Regards,
Lucas
>
> > > Insert an error message to help debug when such an issue happen.
> > >
> > > Signed-off-by: Sui Jingfeng <sui.jingfeng at linux.dev>
> > > ---
> > > For example, when running glmark2-drm:
> > >
> > > [kernel space debug log]
> > >
> > > etnaviv 0000:03:00.0: Insert bo failed, va: fd38b000, size: 4000
> > > etnaviv 0000:03:00.0: Insert bo failed, va: fd38a000, size: 4000
> > >
> > > [user space debug log]
> > >
> > > bo->va = 0xfd48c000, bo->size=100000
> > > bo->va = 0xfd38c000, bo->size=100000
> > > bo->va = 0xfd38b000, bo->size=1000 <-- Insert IOVA fails started at here.
> > > bo->va = 0xfd38a000, bo->size=1000
> > > bo->va = 0xfd389000, bo->size=1000
> > >
> > > [texture] texture-filter=nearest:MESA: error: etna_cmd_stream_flush:238: submit failed: -28 (No space left on device)
> > > ---
> > > drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++-
> > > 1 file changed, 5 insertions(+), 1 deletion(-)
> > >
> > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
> > > index 1661d589bf3e..682f27b27d59 100644
> > > --- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
> > > +++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
> > > @@ -310,8 +310,12 @@ int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context,
> > > else
> > > ret = etnaviv_iommu_find_iova(context, node,
> > > etnaviv_obj->base.size);
> > > - if (ret < 0)
> > > + if (ret < 0) {
> > > + dev_err(context->global->dev,
> > > + "Insert iova failed, va: %llx, size: %zx\n",
> > > + va, etnaviv_obj->base.size);
> > As this might happen for a lot of buffers in a single submit and
> > userspace might be unimpressed by the submit failure and keep pushing
> > new submits, this has a potential to spam the logs. Please use
> > dev_err_ratelimited. Other than that, this patch looks good.
> >
> > Regards,
> > Lucas
> >
> > > goto unlock;
> > > + }
> > >
> > > mapping->iova = node->start;
> > > ret = etnaviv_iommu_map(context, node->start, sgt,
>
More information about the dri-devel
mailing list