[PATCH v2 1/8] drm: Add dummy page per device or GEM object

Christian König christian.koenig at amd.com
Mon Nov 16 09:48:14 UTC 2020


Am 15.11.20 um 07:34 schrieb Andrey Grodzovsky:
>
> On 11/14/20 4:51 AM, Daniel Vetter wrote:
>> On Sat, Nov 14, 2020 at 9:41 AM Christian König
>> <ckoenig.leichtzumerken at gmail.com> wrote:
>>> Am 13.11.20 um 21:52 schrieb Andrey Grodzovsky:
>>>> On 6/22/20 1:50 PM, Daniel Vetter wrote:
>>>>> On Mon, Jun 22, 2020 at 7:45 PM Christian König
>>>>> <christian.koenig at amd.com> wrote:
>>>>>> Am 22.06.20 um 16:32 schrieb Andrey Grodzovsky:
>>>>>>> On 6/22/20 9:18 AM, Christian König wrote:
>>>>>>>> Am 21.06.20 um 08:03 schrieb Andrey Grodzovsky:
>>>>>>>>> Will be used to reroute CPU mapped BO's page faults once
>>>>>>>>> device is removed.
>>>>>>>>>
>>>>>>>>> Signed-off-by: Andrey Grodzovsky <andrey.grodzovsky at amd.com>
>>>>>>>>> ---
>>>>>>>>>     drivers/gpu/drm/drm_file.c  |  8 ++++++++
>>>>>>>>>     drivers/gpu/drm/drm_prime.c | 10 ++++++++++
>>>>>>>>>     include/drm/drm_file.h      |  2 ++
>>>>>>>>>     include/drm/drm_gem.h       |  2 ++
>>>>>>>>>     4 files changed, 22 insertions(+)
>>>>>>>>>
>>>>>>>>> diff --git a/drivers/gpu/drm/drm_file.c 
>>>>>>>>> b/drivers/gpu/drm/drm_file.c
>>>>>>>>> index c4c704e..67c0770 100644
>>>>>>>>> --- a/drivers/gpu/drm/drm_file.c
>>>>>>>>> +++ b/drivers/gpu/drm/drm_file.c
>>>>>>>>> @@ -188,6 +188,12 @@ struct drm_file *drm_file_alloc(struct
>>>>>>>>> drm_minor *minor)
>>>>>>>>>                 goto out_prime_destroy;
>>>>>>>>>         }
>>>>>>>>>     +    file->dummy_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
>>>>>>>>> +    if (!file->dummy_page) {
>>>>>>>>> +        ret = -ENOMEM;
>>>>>>>>> +        goto out_prime_destroy;
>>>>>>>>> +    }
>>>>>>>>> +
>>>>>>>>>         return file;
>>>>>>>>>       out_prime_destroy:
>>>>>>>>> @@ -284,6 +290,8 @@ void drm_file_free(struct drm_file *file)
>>>>>>>>>         if (dev->driver->postclose)
>>>>>>>>>             dev->driver->postclose(dev, file);
>>>>>>>>>     +    __free_page(file->dummy_page);
>>>>>>>>> +
>>>>>>>>> drm_prime_destroy_file_private(&file->prime);
>>>>>>>>> WARN_ON(!list_empty(&file->event_list));
>>>>>>>>> diff --git a/drivers/gpu/drm/drm_prime.c
>>>>>>>>> b/drivers/gpu/drm/drm_prime.c
>>>>>>>>> index 1de2cde..c482e9c 100644
>>>>>>>>> --- a/drivers/gpu/drm/drm_prime.c
>>>>>>>>> +++ b/drivers/gpu/drm/drm_prime.c
>>>>>>>>> @@ -335,6 +335,13 @@ int drm_gem_prime_fd_to_handle(struct
>>>>>>>>> drm_device *dev,
>>>>>>>>>           ret = drm_prime_add_buf_handle(&file_priv->prime,
>>>>>>>>>                 dma_buf, *handle);
>>>>>>>>> +
>>>>>>>>> +    if (!ret) {
>>>>>>>>> +        obj->dummy_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
>>>>>>>>> +        if (!obj->dummy_page)
>>>>>>>>> +            ret = -ENOMEM;
>>>>>>>>> +    }
>>>>>>>>> +
>>>>>>>> While the per file case still looks acceptable this is a clear NAK
>>>>>>>> since it will massively increase the memory needed for a prime
>>>>>>>> exported object.
>>>>>>>>
>>>>>>>> I think that this is quite overkill in the first place and for the
>>>>>>>> hot unplug case we can just use the global dummy page as well.
>>>>>>>>
>>>>>>>> Christian.
>>>>>>> Global dummy page is good for read access, what do you do on write
>>>>>>> access ? My first approach was indeed to map at first global dummy
>>>>>>> page as read only and mark the vma->vm_flags as !VM_SHARED assuming
>>>>>>> that this would trigger Copy On Write flow in core mm
>>>>>>> (https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Fv5.7-rc7%2Fsource%2Fmm%2Fmemory.c%23L3977&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=kghiG3VpCJod6YefExoDVPl9X03zNhw3SN5GAxgbnmU%3D&reserved=0) 
>>>>>>>
>>>>>>>
>>>>>>> on the next page fault to same address triggered by a write 
>>>>>>> access but
>>>>>>> then i realized a new COW page will be allocated for each such 
>>>>>>> mapping
>>>>>>> and this is much more wasteful then having a dedicated page per GEM
>>>>>>> object.
>>>>>> Yeah, but this is only for a very very small corner cases. What 
>>>>>> we need
>>>>>> to prevent is increasing the memory usage during normal operation to
>>>>>> much.
>>>>>>
>>>>>> Using memory during the unplug is completely unproblematic 
>>>>>> because we
>>>>>> just released quite a bunch of it by releasing all those system 
>>>>>> memory
>>>>>> buffers.
>>>>>>
>>>>>> And I'm pretty sure that COWed pages are correctly accounted towards
>>>>>> the
>>>>>> used memory of a process.
>>>>>>
>>>>>> So I think if that approach works as intended and the COW pages are
>>>>>> released again on unmapping it would be the perfect solution to the
>>>>>> problem.
>>>>>>
>>>>>> Daniel what do you think?
>>>>> If COW works, sure sounds reasonable. And if we can make sure we
>>>>> managed to drop all the system allocations (otherwise suddenly 2x
>>>>> memory usage, worst case). But I have no idea whether we can
>>>>> retroshoehorn that into an established vma, you might have fun stuff
>>>>> like a mkwrite handler there (which I thought is the COW handler
>>>>> thing, but really no idea).
>>>>>
>>>>> If we need to massively change stuff then I think rw dummy page,
>>>>> allocated on first fault after hotunplug (maybe just make it one per
>>>>> object, that's simplest) seems like the much safer option. Much less
>>>>> code that can go wrong.
>>>>> -Daniel
>>>>
>>>> Regarding COW, i was looking into how to properly implement it from
>>>> within the fault handler (i.e. ttm_bo_vm_fault)
>>>> and the main obstacle I hit is that of exclusive access to the
>>>> vm_area_struct, i need to be able to modify
>>>> vma->vm_flags (and vm_page_prot)  to remove VM_SHARED bit so COW can
>>>> be triggered on subsequent write access
>>>> fault (here
>>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Flatest%2Fsource%2Fmm%2Fmemory.c%23L4128&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=ziHJtqyHuLrlb0uYKhoWCWhUAZnX0JquE%2BkBJ5Fx%2BNo%3D&reserved=0) 
>>>>
>>>> but core mm takes only read side mm_sem (here for example
>>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Flatest%2Fsource%2Fdrivers%2Fiommu%2Famd%2Fiommu_v2.c%23L488&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=h360c75Upl3%2FW7im7M1%2BxY%2FXy4gxin%2BkCF1Ui2zFXMs%3D&reserved=0) 
>>>>
>>>> and so I am not supposed to modify vm_area_struct in this case. I am
>>>> not sure if it's legit to write lock tthe mm_sem from this point.
>>>> I found some discussions about this here
>>>> https://nam11.safelinks.protection.outlook.com/?url=http%3A%2F%2Flkml.iu.edu%2Fhypermail%2Flinux%2Fkernel%2F1909.1%2F02754.html&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224021379%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=sx6s1lH%2FvxbIZajc4Yr49vFhxvPEnBHZlTt52D8qvZA%3D&reserved=0 
>>>> but it
>>>> wasn't really clear to me
>>>> what's the solution.
>>>>
>>>> In any case, seems to me that easier and more memory saving solution
>>>> would be to just switch to per ttm bo dumy rw page that
>>>> would be allocated on demand as you suggested here.  This should also
>>>> take care of imported BOs and flink cases.
>>>> Then i can drop the per device FD and per GEM object FD dummy BO and
>>>> the ugly loop i am using in patch 2 to match faulting BO to the right
>>>> dummy page.
>>>>
>>>> Does this makes sense ?
>>> I still don't see the information leak as much of a problem, but if
>>> Daniel insists we should probably do this.
>> Well amdgpu doesn't clear buffers by default, so indeed you guys are a
>> lot more laissez-faire here. But in general we really don't do that
>> kind of leaking. Iirc there's even radeonsi bugs because else clears,
>> and radeonsi happily displays gunk :-)
>>
>>> But could we at least have only one page per client instead of per BO?
>> I think you can do one page per file descriptor or something like
>> that. But gets annoying with shared bo, especially with dma_buf_mmap
>> forwarding.
>> -Daniel
>
>
> Christian - is your concern more with too much page allocations or 
> with extra pointer member
> cluttering TTM BO struct ?

Yes, that is one problem.

> Because we can allocate the dummy page on demand only when
> needed. It's just seems to me that keeping it per BO streamlines the 
> code as I don't need to
> have different handling for local vs imported BOs.

Why should you have a difference between local vs imported BOs?

Christian.

>
> Andrey



More information about the amd-gfx mailing list