[PATCH v4 1/5] udmabuf: direct map pfn when first page fault
Huan Yang
link at vivo.com
Thu Aug 29 06:35:48 UTC 2024
在 2024/8/29 14:08, Kasireddy, Vivek 写道:
> Hi Huan,
>
>> Subject: [PATCH v4 1/5] udmabuf: direct map pfn when first page fault
>>
>> The current udmabuf mmap uses a page fault to populate the vma.
>>
>> However, the current udmabuf has already obtained and pinned the folio
>> upon completion of the creation.This means that the physical memory has
>> already been acquired, rather than being accessed dynamically.
>>
>> As a result, the page fault has lost its purpose as a demanding
>> page. Due to the fact that page fault requires trapping into kernel mode
>> and filling in when accessing the corresponding virtual address in mmap,
>> when creating a large size udmabuf, this represents a considerable
>> overhead.
>>
>> This patch fill vma area with pfn when the first page fault trigger, so,
>> any other access will not enter page fault.
>>
>> Suggested-by: Vivek Kasireddy <vivek.kasireddy at intel.com>
>> Signed-off-by: Huan Yang <link at vivo.com>
>> ---
>> drivers/dma-buf/udmabuf.c | 26 ++++++++++++++++++++++++--
>> 1 file changed, 24 insertions(+), 2 deletions(-)
>>
>> diff --git a/drivers/dma-buf/udmabuf.c b/drivers/dma-buf/udmabuf.c
>> index 047c3cd2ceff..0e33d25310ec 100644
>> --- a/drivers/dma-buf/udmabuf.c
>> +++ b/drivers/dma-buf/udmabuf.c
>> @@ -43,7 +43,8 @@ static vm_fault_t udmabuf_vm_fault(struct vm_fault
>> *vmf)
>> struct vm_area_struct *vma = vmf->vma;
>> struct udmabuf *ubuf = vma->vm_private_data;
>> pgoff_t pgoff = vmf->pgoff;
>> - unsigned long pfn;
>> + unsigned long addr, end, pfn;
>> + vm_fault_t ret;
>>
>> if (pgoff >= ubuf->pagecount)
>> return VM_FAULT_SIGBUS;
>> @@ -51,7 +52,28 @@ static vm_fault_t udmabuf_vm_fault(struct vm_fault
>> *vmf)
>> pfn = folio_pfn(ubuf->folios[pgoff]);
>> pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT;
>>
>> - return vmf_insert_pfn(vma, vmf->address, pfn);
>> + ret = vmf_insert_pfn(vma, vmf->address, pfn);
>> + if (ret & VM_FAULT_ERROR)
>> + return ret;
>> +
>> + /* pre fault */
>> + pgoff = vma->vm_pgoff;
>> + end = vma->vm_end;
>> + addr = vma->vm_start;
>> +
>> + for (; addr < end; pgoff++, addr += PAGE_SIZE) {
> Although unlikely, I think we should also check for pgoff < ubuf->pagecount.
Ohh, yes.
>
>> + if (addr == vmf->address)
>> + continue;
>> +
>> + pfn = folio_pfn(ubuf->folios[pgoff]);
>> +
>> + pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT;
>> +
>> + if (vmf_insert_pfn(vma, addr, pfn) & VM_FAULT_ERROR)
> Shouldn't you store the return value of vmf_insert_pfn in ret? Otherwise, we'll
> return success when the above call fails.
If anything wrong in there, let it report when it true access and then
report. Not in there.
It's pre-fault, 'this' page's fault is already dealed success. No reason
to report in it's fault.
>
> Anyway, I am wondering if it is more optimal to just iterate over pages instead
> of addresses. Something like below:
>
> + unsigned long nr_pages = vma_pages(vma);
> + unsigned long addr = vma->vm_start;
>
> - if (pgoff >= ubuf->pagecount)
> - return VM_FAULT_SIGBUS;
> + WARN_ON(nr_pages != ubuf->pagecount);
>
> - pfn = folio_pfn(ubuf->folios[pgoff]);
> - pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT;
> + for (pg = 0; pg < nr_pages && pg < ubuf->pagecount; pg++) {
> + pfn = folio_pfn(ubuf->folios[pg]);
> + pfn += ubuf->offsets[pg] >> PAGE_SHIFT;
>
> - return vmf_insert_pfn(vma, vmf->address, pfn);
> + ret = vmf_insert_pfn(vma, addr, pfn);
> + addr += PAGE_SIZE;
> + }
Maybe mmap with a offset? just start from 0 not too good.
This seems clear, but is it not good to handle insert pfn error elsewhere?
>
> Thanks,
> Vivek
>
>> + break;
>> + }
>> +
>> + return ret;
>> }
>>
>> static const struct vm_operations_struct udmabuf_vm_ops = {
>> --
>> 2.45.2
More information about the dri-devel
mailing list