[PATCH 2/2] dmabuf/heaps: implement DMA_BUF_IOCTL_RW_FILE for system_heap

wangtao tao.wangtao at honor.com
Wed May 14 11:02:09 UTC 2025



> -----Original Message-----
> From: Christian König <christian.koenig at amd.com>
> Sent: Tuesday, May 13, 2025 9:18 PM
> To: wangtao <tao.wangtao at honor.com>; sumit.semwal at linaro.org;
> benjamin.gaignard at collabora.com; Brian.Starkey at arm.com;
> jstultz at google.com; tjmercier at google.com
> Cc: linux-media at vger.kernel.org; dri-devel at lists.freedesktop.org; linaro-
> mm-sig at lists.linaro.org; linux-kernel at vger.kernel.org;
> wangbintian(BintianWang) <bintian.wang at honor.com>; yipengxiang
> <yipengxiang at honor.com>; <liulu.liu at honor.com>; <feng.han at honor.com>
> Subject: Re: [PATCH 2/2] dmabuf/heaps: implement
> DMA_BUF_IOCTL_RW_FILE for system_heap
> 
> On 5/13/25 14:30, wangtao wrote:
> >> -----Original Message-----
> >> From: Christian König <christian.koenig at amd.com>
> >> Sent: Tuesday, May 13, 2025 7:32 PM
> >> To: wangtao <tao.wangtao at honor.com>; sumit.semwal at linaro.org;
> >> benjamin.gaignard at collabora.com; Brian.Starkey at arm.com;
> >> jstultz at google.com; tjmercier at google.com
> >> Cc: linux-media at vger.kernel.org; dri-devel at lists.freedesktop.org;
> >> linaro- mm-sig at lists.linaro.org; linux-kernel at vger.kernel.org;
> >> wangbintian(BintianWang) <bintian.wang at honor.com>; yipengxiang
> >> <yipengxiang at honor.com>; <liulu.liu at honor.com>;
> >> <feng.han at honor.com>
> >> Subject: Re: [PATCH 2/2] dmabuf/heaps: implement
> >> DMA_BUF_IOCTL_RW_FILE for system_heap
> >>
> >> On 5/13/25 11:28, wangtao wrote:
> >>> Support direct file I/O operations for system_heap dma-buf objects.
> >>> Implementation includes:
> >>> 1. Convert sg_table to bio_vec
> >>
> >> That is usually illegal for DMA-bufs.
> > [wangtao] The term 'convert' is misleading in this context. The appropriate
> phrasing should be: Construct bio_vec from sg_table.
> 
> Well it doesn't matter what you call it. Touching the page inside an sg table of
> a DMA-buf is illegal, we even have code to actively prevent that.
[wangtao] For a driver using DMA-buf: Don't touch pages in the sg_table. But the system heap exporter (sg_table owner) should be allowed to use them.
If a driver takes ownership via dma_buf_map_attachment or similar calls, the exporter must stop using the sg_table.
User-space programs should call DMA_BUF_IOCTL_RW_FILE only when the DMA-buf is not attached.
The exporter must check ownership (e.g., ensure no map_dma_buf/vmap is active) and block new calls during operations.
I'll add these checks in patch v2.

> 
> Once more: This approach was already rejected multiple times! Please use
> udmabuf instead!
> 
> The hack you came up here is simply not necessary.
[wangtao] Many people need DMA-buf direct I/O. I tried it 2 years ago. My method is simpler, uses less CPU/power, and performs better:
  - Speed: 3418 MB/s vs. 2073 MB/s (udmabuf) at 1GHz CPU.
  - udmabuf wastes half its CPU time on __get_user_pages.
  - Creating 32x32MB DMA-bufs + reading 1GB file takes 346 ms vs. 1145 ms for udmabuf (10x slower) vs. 1503 ms for DMA-buf normal.
udmabuf is slightly faster but not enough. Switching to udmabuf is easy for small apps but hard in complex systems without major benefits.
> 
> Regards,
> Christian.
> 
> 
> > Appreciate your feedback.
> >>
> >> Regards,
> >> Christian.
> >>
> >>> 2. Set IOCB_DIRECT when O_DIRECT is supported 3. Invoke
> >>> vfs_iocb_iter_read()/vfs_iocb_iter_write() for actual I/O
> >>>
> >>> Performance metrics (UFS 4.0 device @4GB/s, Arm64 CPU @1GHz):
> >>>
> >>> | Metric             |    1MB |    8MB |    64MB |   1024MB |   3072MB |
> >>> |--------------------|-------:|-------:|--------:|---------:|-------
> >>> |--------------------|--
> >>> |--------------------|:|
> >>> | Buffer Read (us)   |   1658 |   9028 |   69295 |  1019783 |  2978179 |
> >>> | Direct Read (us)   |    707 |   2647 |   18689 |   299627 |   937758 |
> >>> | Buffer Rate (MB/s) |    603 |    886 |     924 |     1004 |     1032 |
> >>> | Direct Rate (MB/s) |   1414 |   3022 |    3425 |     3418 |     3276 |
> >>>
> >>> Signed-off-by: wangtao <tao.wangtao at honor.com>
> >>> ---
> >>>  drivers/dma-buf/heaps/system_heap.c | 118
> >>> ++++++++++++++++++++++++++++
> >>>  1 file changed, 118 insertions(+)
> >>>
> >>> diff --git a/drivers/dma-buf/heaps/system_heap.c
> >>> b/drivers/dma-buf/heaps/system_heap.c
> >>> index 26d5dc89ea16..f7b71b9843aa 100644
> >>> --- a/drivers/dma-buf/heaps/system_heap.c
> >>> +++ b/drivers/dma-buf/heaps/system_heap.c
> >>> @@ -20,6 +20,8 @@
> >>>  #include <linux/scatterlist.h>
> >>>  #include <linux/slab.h>
> >>>  #include <linux/vmalloc.h>
> >>> +#include <linux/bvec.h>
> >>> +#include <linux/uio.h>
> >>>
> >>>  static struct dma_heap *sys_heap;
> >>>
> >>> @@ -281,6 +283,121 @@ static void system_heap_vunmap(struct
> dma_buf
> >> *dmabuf, struct iosys_map *map)
> >>>  	iosys_map_clear(map);
> >>>  }
> >>>
> >>> +static struct bio_vec *system_heap_init_bvec(struct
> >> system_heap_buffer *buffer,
> >>> +			size_t offset, size_t len, int *nr_segs) {
> >>> +	struct sg_table *sgt = &buffer->sg_table;
> >>> +	struct scatterlist *sg;
> >>> +	size_t length = 0;
> >>> +	unsigned int i, k = 0;
> >>> +	struct bio_vec *bvec;
> >>> +	size_t sg_left;
> >>> +	size_t sg_offset;
> >>> +	size_t sg_len;
> >>> +
> >>> +	bvec = kvcalloc(sgt->nents, sizeof(*bvec), GFP_KERNEL);
> >>> +	if (!bvec)
> >>> +		return NULL;
> >>> +
> >>> +	for_each_sg(sgt->sgl, sg, sgt->nents, i) {
> >>> +		length += sg->length;
> >>> +		if (length <= offset)
> >>> +			continue;
> >>> +
> >>> +		sg_left = length - offset;
> >>> +		sg_offset = sg->offset + sg->length - sg_left;
> >>> +		sg_len = min(sg_left, len);
> >>> +
> >>> +		bvec[k].bv_page = sg_page(sg);
> >>> +		bvec[k].bv_len = sg_len;
> >>> +		bvec[k].bv_offset = sg_offset;
> >>> +		k++;
> >>> +
> >>> +		offset += sg_len;
> >>> +		len -= sg_len;
> >>> +		if (len <= 0)
> >>> +			break;
> >>> +	}
> >>> +
> >>> +	*nr_segs = k;
> >>> +	return bvec;
> >>> +}
> >>> +
> >>> +static int system_heap_rw_file(struct system_heap_buffer *buffer,
> >>> +bool
> >> is_read,
> >>> +		bool direct_io, struct file *filp, loff_t file_offset,
> >>> +		size_t buf_offset, size_t len)
> >>> +{
> >>> +	struct bio_vec *bvec;
> >>> +	int nr_segs = 0;
> >>> +	struct iov_iter iter;
> >>> +	struct kiocb kiocb;
> >>> +	ssize_t ret = 0;
> >>> +
> >>> +	if (direct_io) {
> >>> +		if (!(filp->f_mode & FMODE_CAN_ODIRECT))
> >>> +			return -EINVAL;
> >>> +	}
> >>> +
> >>> +	bvec = system_heap_init_bvec(buffer, buf_offset, len, &nr_segs);
> >>> +	if (!bvec)
> >>> +		return -ENOMEM;
> >>> +
> >>> +	iov_iter_bvec(&iter, is_read ? ITER_DEST : ITER_SOURCE, bvec,
> >> nr_segs, len);
> >>> +	init_sync_kiocb(&kiocb, filp);
> >>> +	kiocb.ki_pos = file_offset;
> >>> +	if (direct_io)
> >>> +		kiocb.ki_flags |= IOCB_DIRECT;
> >>> +
> >>> +	while (kiocb.ki_pos < file_offset + len) {
> >>> +		if (is_read)
> >>> +			ret = vfs_iocb_iter_read(filp, &kiocb, &iter);
> >>> +		else
> >>> +			ret = vfs_iocb_iter_write(filp, &kiocb, &iter);
> >>> +		if (ret <= 0)
> >>> +			break;
> >>> +	}
> >>> +
> >>> +	kvfree(bvec);
> >>> +	return ret < 0 ? ret : 0;
> >>> +}
> >>> +
> >>> +static int system_heap_dma_buf_rw_file(struct dma_buf *dmabuf,
> >>> +			struct dma_buf_rw_file *back)
> >>> +{
> >>> +	struct system_heap_buffer *buffer = dmabuf->priv;
> >>> +	int ret = 0;
> >>> +	__u32 op = back->flags & DMA_BUF_RW_FLAGS_OP_MASK;
> >>> +	bool direct_io = back->flags & DMA_BUF_RW_FLAGS_DIRECT;
> >>> +	struct file *filp;
> >>> +
> >>> +	if (op != DMA_BUF_RW_FLAGS_READ && op !=
> >> DMA_BUF_RW_FLAGS_WRITE)
> >>> +		return -EINVAL;
> >>> +	if (direct_io) {
> >>> +		if (!PAGE_ALIGNED(back->file_offset) ||
> >>> +			!PAGE_ALIGNED(back->buf_offset) ||
> >>> +			!PAGE_ALIGNED(back->buf_len))
> >>> +		return -EINVAL;
> >>> +	}
> >>> +	if (!back->buf_len || back->buf_len > dmabuf->size ||
> >>> +		back->buf_offset >= dmabuf->size ||
> >>> +		back->buf_offset + back->buf_len > dmabuf->size)
> >>> +		return -EINVAL;
> >>> +	if (back->file_offset + back->buf_len < back->file_offset)
> >>> +		return -EINVAL;
> >>> +
> >>> +	filp = fget(back->fd);
> >>> +	if (!filp)
> >>> +		return -EBADF;
> >>> +
> >>> +	mutex_lock(&buffer->lock);
> >>> +	ret = system_heap_rw_file(buffer, op ==
> >> DMA_BUF_RW_FLAGS_READ, direct_io,
> >>> +			filp, back->file_offset, back->buf_offset, back-
> >>> buf_len);
> >>> +	mutex_unlock(&buffer->lock);
> >>> +
> >>> +	fput(filp);
> >>> +	return ret;
> >>> +}
> >>> +
> >>>  static void system_heap_dma_buf_release(struct dma_buf *dmabuf)  {
> >>>  	struct system_heap_buffer *buffer = dmabuf->priv; @@ -308,6
> >> +425,7
> >>> @@ static const struct dma_buf_ops system_heap_buf_ops = {
> >>>  	.mmap = system_heap_mmap,
> >>>  	.vmap = system_heap_vmap,
> >>>  	.vunmap = system_heap_vunmap,
> >>> +	.rw_file = system_heap_dma_buf_rw_file,
> >>>  	.release = system_heap_dma_buf_release,  };
> >>>
> >



More information about the dri-devel mailing list