[PATCH v12 1/8] mm/gup: Introduce unpin_folio/unpin_folios helpers
David Hildenbrand
david at redhat.com
Tue Apr 2 14:14:09 UTC 2024
On 02.04.24 15:52, David Hildenbrand wrote:
> On 25.02.24 08:56, Vivek Kasireddy wrote:
>> These helpers are the folio versions of unpin_user_page/unpin_user_pages.
>> They are currently only useful for unpinning folios pinned by
>> memfd_pin_folios() or other associated routines. However, they could
>> find new uses in the future, when more and more folio-only helpers
>> are added to GUP.
>>
>> Cc: David Hildenbrand <david at redhat.com>
>> Cc: Matthew Wilcox <willy at infradead.org>
>> Cc: Christoph Hellwig <hch at infradead.org>
>> Cc: Jason Gunthorpe <jgg at nvidia.com>
>> Cc: Peter Xu <peterx at redhat.com>
>> Suggested-by: David Hildenbrand <david at redhat.com>
>> Signed-off-by: Vivek Kasireddy <vivek.kasireddy at intel.com>
>> ---
>> include/linux/mm.h | 2 ++
>> mm/gup.c | 81 ++++++++++++++++++++++++++++++++++++++++------
>> 2 files changed, 74 insertions(+), 9 deletions(-)
>>
>> diff --git a/include/linux/mm.h b/include/linux/mm.h
>> index 6f4825d82965..36e4c2b22600 100644
>> --- a/include/linux/mm.h
>> +++ b/include/linux/mm.h
>> @@ -1601,11 +1601,13 @@ static inline void put_page(struct page *page)
>> #define GUP_PIN_COUNTING_BIAS (1U << 10)
>>
>> void unpin_user_page(struct page *page);
>> +void unpin_folio(struct folio *folio);
>> void unpin_user_pages_dirty_lock(struct page **pages, unsigned long npages,
>> bool make_dirty);
>> void unpin_user_page_range_dirty_lock(struct page *page, unsigned long npages,
>> bool make_dirty);
>> void unpin_user_pages(struct page **pages, unsigned long npages);
>> +void unpin_folios(struct folio **folios, unsigned long nfolios);
>>
>> static inline bool is_cow_mapping(vm_flags_t flags)
>> {
>> diff --git a/mm/gup.c b/mm/gup.c
>> index df83182ec72d..0a45eda6aaeb 100644
>> --- a/mm/gup.c
>> +++ b/mm/gup.c
>> @@ -30,6 +30,23 @@ struct follow_page_context {
>> unsigned int page_mask;
>> };
>>
>> +static inline void sanity_check_pinned_folios(struct folio **folios,
>> + unsigned long nfolios)
>> +{
>> + if (!IS_ENABLED(CONFIG_DEBUG_VM))
>> + return;
>> +
>> + for (; nfolios; nfolios--, folios++) {
>> + struct folio *folio = *folios;
>> +
>> + if (is_zero_folio(folio) ||
>> + !folio_test_anon(folio))
>> + continue;
>> +
>> + VM_BUG_ON_FOLIO(!PageAnonExclusive(&folio->page), folio);
>
> That change is wrong (and the split makes the check confusing).
>
> It could be that the first subpage is no longer exclusive, but the given
> (sanity_check_pinned_pages() ) subpage is exclusive for large folios.
>
> I suggest dropping that change, and instead, in
> unpin_folio()/unpin_folios(), reject any anon folios for now.
>
> So, replace the sanity_check_pinned_folios() in unpin_folio() /
> unpin_folios() by a VM_WARN_ON(folio_test_anon(folio));
After reading patch #2: drop both the sanity check and VM_WARN_ON() from
unpin_folio()/unpin_folios(), and add a comment to the patch description
that we cannot do the sanity checking without the subpage, and that we
can reintroduce it once we have a single per-folio AnonExclusive bit.
--
Cheers,
David / dhildenb
More information about the dri-devel
mailing list