[PATCH v4 0/4] Implement dmabuf direct I/O via copy_file_range
wangtao
tao.wangtao at honor.com
Fri Jun 6 09:52:54 UTC 2025
> -----Original Message-----
> From: Christoph Hellwig <hch at infradead.org>
> Sent: Tuesday, June 3, 2025 9:20 PM
> To: Christian König <christian.koenig at amd.com>
> Cc: Christoph Hellwig <hch at infradead.org>; wangtao
> <tao.wangtao at honor.com>; sumit.semwal at linaro.org; kraxel at redhat.com;
> vivek.kasireddy at intel.com; viro at zeniv.linux.org.uk; brauner at kernel.org;
> hughd at google.com; akpm at linux-foundation.org; amir73il at gmail.com;
> benjamin.gaignard at collabora.com; Brian.Starkey at arm.com;
> jstultz at google.com; tjmercier at google.com; jack at suse.cz;
> baolin.wang at linux.alibaba.com; linux-media at vger.kernel.org; dri-
> devel at lists.freedesktop.org; linaro-mm-sig at lists.linaro.org; linux-
> kernel at vger.kernel.org; linux-fsdevel at vger.kernel.org; linux-
> mm at kvack.org; wangbintian(BintianWang) <bintian.wang at honor.com>;
> yipengxiang <yipengxiang at honor.com>; liulu 00013167
> <liulu.liu at honor.com>; hanfeng 00012985 <feng.han at honor.com>
> Subject: Re: [PATCH v4 0/4] Implement dmabuf direct I/O via
> copy_file_range
>
> On Tue, Jun 03, 2025 at 03:14:20PM +0200, Christian König wrote:
> > On 6/3/25 15:00, Christoph Hellwig wrote:
> > > This is a really weird interface. No one has yet to explain why
> > > dmabuf is so special that we can't support direct I/O to it when we
> > > can support it to otherwise exotic mappings like PCI P2P ones.
> >
> > With udmabuf you can do direct I/O, it's just inefficient to walk the
> > page tables for it when you already have an array of all the folios.
>
> Does it matter compared to the I/O in this case?
>
> Either way there has been talk (in case of networking implementations) that
> use a dmabuf as a first class container for lower level I/O.
> I'd much rather do that than adding odd side interfaces. I.e. have a version
> of splice that doesn't bother with the pipe, but instead just uses in-kernel
> direct I/O on one side and dmabuf-provided folios on the other.
If the VFS layer recognizes dmabuf type and acquires its sg_table
and folios, zero-copy could also be achieved. I initially thought
dmabuf acts as a driver and shouldn't be handled by VFS, so I made
dmabuf implement copy_file_range callbacks to support direct I/O
zero-copy. I'm open to both approaches. What's the preference of
VFS experts?
Regards,
Wangtao.
More information about the dri-devel
mailing list