[Nouveau] [PATCH 4/5] drm/ttm: add ttm_sg_tt_init
Daniel Vetter
daniel at ffwll.ch
Tue Mar 6 09:19:07 UTC 2018
On Tue, Feb 27, 2018 at 01:07:06PM +0100, Christian König wrote:
> Hi guys,
>
> at least on amdgpu and radeon the page array allocated by ttm_dma_tt_init is
> completely unused in the case of DMA-buf sharing. So I'm trying to get rid
> of that by only allocating the DMA address array.
>
> Now the only other user of DMA-buf together with ttm_dma_tt_init is Nouveau.
> So my question is are you guys using the page array anywhere in your kernel
> driver in case of a DMA-buf sharing?
>
> If no then I could just make this the default behavior for all drivers and
> save quite a bit of memory for everybody.
+1 on teaching ttm to no longer look at the struct page * in the dma-buf
sgt, but only the dma_buf address.
If there's still some need for in-kernel cpu or userspace mmap access then
imo ttm needs to be fixed to delegate all that to the right dma-buf
interfaces. The ttm abstraction is already there, it's just not passed
through.
I don't pretend to now enough of the details to review this stuff :-)
-Daniel
>
> Thanks,
> Christian.
>
> Am 27.02.2018 um 12:49 schrieb Christian König:
> > This allows drivers to only allocate dma addresses, but not a page
> > array.
> >
> > Signed-off-by: Christian König <christian.koenig at amd.com>
> > ---
> > drivers/gpu/drm/ttm/ttm_tt.c | 54 ++++++++++++++++++++++++++++++++++++--------
> > include/drm/ttm/ttm_tt.h | 2 ++
> > 2 files changed, 47 insertions(+), 9 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c
> > index 8e0b525cda00..971133106ec2 100644
> > --- a/drivers/gpu/drm/ttm/ttm_tt.c
> > +++ b/drivers/gpu/drm/ttm/ttm_tt.c
> > @@ -108,6 +108,16 @@ static int ttm_dma_tt_alloc_page_directory(struct ttm_dma_tt *ttm)
> > return 0;
> > }
> > +static int ttm_sg_tt_alloc_page_directory(struct ttm_dma_tt *ttm)
> > +{
> > + ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages,
> > + sizeof(*ttm->dma_address),
> > + GFP_KERNEL | __GFP_ZERO);
> > + if (!ttm->dma_address)
> > + return -ENOMEM;
> > + return 0;
> > +}
> > +
> > #ifdef CONFIG_X86
> > static inline int ttm_tt_set_page_caching(struct page *p,
> > enum ttm_caching_state c_old,
> > @@ -227,8 +237,8 @@ void ttm_tt_destroy(struct ttm_tt *ttm)
> > ttm->func->destroy(ttm);
> > }
> > -int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
> > - unsigned long size, uint32_t page_flags)
> > +void ttm_tt_init_fields(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
> > + unsigned long size, uint32_t page_flags)
> > {
> > ttm->bdev = bdev;
> > ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
> > @@ -236,6 +246,12 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
> > ttm->page_flags = page_flags;
> > ttm->state = tt_unpopulated;
> > ttm->swap_storage = NULL;
> > +}
> > +
> > +int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
> > + unsigned long size, uint32_t page_flags)
> > +{
> > + ttm_tt_init_fields(ttm, bdev, size, page_flags);
> > if (ttm_tt_alloc_page_directory(ttm)) {
> > ttm_tt_destroy(ttm);
> > @@ -258,12 +274,7 @@ int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
> > {
> > struct ttm_tt *ttm = &ttm_dma->ttm;
> > - ttm->bdev = bdev;
> > - ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
> > - ttm->caching_state = tt_cached;
> > - ttm->page_flags = page_flags;
> > - ttm->state = tt_unpopulated;
> > - ttm->swap_storage = NULL;
> > + ttm_tt_init_fields(ttm, bdev, size, page_flags);
> > INIT_LIST_HEAD(&ttm_dma->pages_list);
> > if (ttm_dma_tt_alloc_page_directory(ttm_dma)) {
> > @@ -275,11 +286,36 @@ int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
> > }
> > EXPORT_SYMBOL(ttm_dma_tt_init);
> > +int ttm_sg_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
> > + unsigned long size, uint32_t page_flags)
> > +{
> > + struct ttm_tt *ttm = &ttm_dma->ttm;
> > + int ret;
> > +
> > + ttm_tt_init_fields(ttm, bdev, size, page_flags);
> > +
> > + INIT_LIST_HEAD(&ttm_dma->pages_list);
> > + if (page_flags & TTM_PAGE_FLAG_SG)
> > + ret = ttm_sg_tt_alloc_page_directory(ttm_dma);
> > + else
> > + ret = ttm_dma_tt_alloc_page_directory(ttm_dma);
> > + if (ret) {
> > + ttm_tt_destroy(ttm);
> > + pr_err("Failed allocating page table\n");
> > + return -ENOMEM;
> > + }
> > + return 0;
> > +}
> > +EXPORT_SYMBOL(ttm_sg_tt_init);
> > +
> > void ttm_dma_tt_fini(struct ttm_dma_tt *ttm_dma)
> > {
> > struct ttm_tt *ttm = &ttm_dma->ttm;
> > - kvfree(ttm->pages);
> > + if (ttm->pages)
> > + kvfree(ttm->pages);
> > + else
> > + kvfree(ttm_dma->dma_address);
> > ttm->pages = NULL;
> > ttm_dma->dma_address = NULL;
> > }
> > diff --git a/include/drm/ttm/ttm_tt.h b/include/drm/ttm/ttm_tt.h
> > index 9c78556b488e..1cf316a4257c 100644
> > --- a/include/drm/ttm/ttm_tt.h
> > +++ b/include/drm/ttm/ttm_tt.h
> > @@ -163,6 +163,8 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
> > unsigned long size, uint32_t page_flags);
> > int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
> > unsigned long size, uint32_t page_flags);
> > +int ttm_sg_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
> > + unsigned long size, uint32_t page_flags);
> > /**
> > * ttm_tt_fini
>
> _______________________________________________
> Nouveau mailing list
> Nouveau at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/nouveau
--
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
More information about the Nouveau
mailing list