[PATCH 6/8] drm: Add a drm_get_unmapped_area() helper
Christian König
christian.koenig at amd.com
Wed Dec 4 12:08:06 UTC 2019
Am 04.12.19 um 12:36 schrieb Thomas Hellström (VMware):
> On 12/4/19 12:11 PM, Christian König wrote:
>> Am 03.12.19 um 14:22 schrieb Thomas Hellström (VMware):
>>> From: Thomas Hellstrom <thellstrom at vmware.com>
>>>
>>> This helper is used to align user-space buffer object addresses to
>>> huge page boundaries, minimizing the chance of alignment mismatch
>>> between user-space addresses and physical addresses.
>>
>> Mhm, I'm wondering if that is really such a good idea.
>
> Could you elaborate? What drawbacks do you see?
Main problem for me seems to be that I don't fully understand what the
get_unmapped_area callback is doing.
For example why do we need to use drm_vma_offset_lookup_locked() to
adjust the pgoff?
The mapped offset should be completely irrelevant for finding some piece
of userspace address space or am I totally off here?
> Note that this is the way other subsystems are doing it. Take a look
> at shmem_get_unmapped_area() for instance.
>
>>
>> Wouldn't it be sufficient if userspace uses MAP_HUGETLB?
>
> MAP_HUGETLB is something different and appears to be tied to the
> kernel persistent huge page mechanism, whereas the TTM huge pages is
> tided to the THP functionality (although skipped by khugepaged).
Ok, that makes sense. Over all we want to be transparent here.
Regards,
Christian.
>
> Thanks,
>
> Thomas
>
>
>
>>
>> Regards,
>> Christian.
>>
>>>
>>> Cc: Andrew Morton <akpm at linux-foundation.org>
>>> Cc: Michal Hocko <mhocko at suse.com>
>>> Cc: "Matthew Wilcox (Oracle)" <willy at infradead.org>
>>> Cc: "Kirill A. Shutemov" <kirill.shutemov at linux.intel.com>
>>> Cc: Ralph Campbell <rcampbell at nvidia.com>
>>> Cc: "Jérôme Glisse" <jglisse at redhat.com>
>>> Cc: "Christian König" <christian.koenig at amd.com>
>>> Signed-off-by: Thomas Hellstrom <thellstrom at vmware.com>
>>> ---
>>> Â drivers/gpu/drm/drm_file.c | 130
>>> +++++++++++++++++++++++++++++++++++++
>>>  include/drm/drm_file.h    |  5 ++
>>> Â 2 files changed, 135 insertions(+)
>>>
>>> diff --git a/drivers/gpu/drm/drm_file.c b/drivers/gpu/drm/drm_file.c
>>> index ea34bc991858..e5b4024cd397 100644
>>> --- a/drivers/gpu/drm/drm_file.c
>>> +++ b/drivers/gpu/drm/drm_file.c
>>> @@ -31,6 +31,8 @@
>>> Â Â * OTHER DEALINGS IN THE SOFTWARE.
>>> Â Â */
>>> Â +#include <uapi/asm/mman.h>
>>> +
>>> Â #include <linux/dma-fence.h>
>>> Â #include <linux/module.h>
>>> Â #include <linux/pci.h>
>>> @@ -41,6 +43,7 @@
>>> Â #include <drm/drm_drv.h>
>>> Â #include <drm/drm_file.h>
>>> Â #include <drm/drm_print.h>
>>> +#include <drm/drm_vma_manager.h>
>>> Â Â #include "drm_crtc_internal.h"
>>> Â #include "drm_internal.h"
>>> @@ -754,3 +757,130 @@ void drm_send_event(struct drm_device *dev,
>>> struct drm_pending_event *e)
>>> Â Â Â Â Â spin_unlock_irqrestore(&dev->event_lock, irqflags);
>>> Â }
>>> Â EXPORT_SYMBOL(drm_send_event);
>>> +
>>> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE
>>> +/*
>>> + * drm_addr_inflate() attempts to construct an aligned area by
>>> inflating
>>> + * the area size and skipping the unaligned start of the area.
>>> + * adapted from shmem_get_unmapped_area()
>>> + */
>>> +static unsigned long drm_addr_inflate(unsigned long addr,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long len,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long pgoff,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long flags,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long huge_size)
>>> +{
>>> +Â Â Â unsigned long offset, inflated_len;
>>> +Â Â Â unsigned long inflated_addr;
>>> +Â Â Â unsigned long inflated_offset;
>>> +
>>> +Â Â Â offset = (pgoff << PAGE_SHIFT) & (huge_size - 1);
>>> +Â Â Â if (offset && offset + len < 2 * huge_size)
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if ((addr & (huge_size - 1)) == offset)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â inflated_len = len + huge_size - PAGE_SIZE;
>>> +Â Â Â if (inflated_len > TASK_SIZE)
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if (inflated_len < len)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â inflated_addr = current->mm->get_unmapped_area(NULL, 0,
>>> inflated_len,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â 0, flags);
>>> +Â Â Â if (IS_ERR_VALUE(inflated_addr))
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if (inflated_addr & ~PAGE_MASK)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â inflated_offset = inflated_addr & (huge_size - 1);
>>> +Â Â Â inflated_addr += offset - inflated_offset;
>>> +Â Â Â if (inflated_offset > offset)
>>> +Â Â Â Â Â Â Â inflated_addr += huge_size;
>>> +
>>> +Â Â Â if (inflated_addr > TASK_SIZE - len)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â return inflated_addr;
>>> +}
>>> +
>>> +/**
>>> + * drm_get_unmapped_area() - Get an unused user-space virtual
>>> memory area
>>> + * suitable for huge page table entries.
>>> + * @file: The struct file representing the address space being
>>> mmap()'d.
>>> + * @uaddr: Start address suggested by user-space.
>>> + * @len: Length of the area.
>>> + * @pgoff: The page offset into the address space.
>>> + * @flags: mmap flags
>>> + * @mgr: The address space manager used by the drm driver. This
>>> argument can
>>> + * probably be removed at some point when all drivers use the same
>>> + * address space manager.
>>> + *
>>> + * This function attempts to find an unused user-space virtual
>>> memory area
>>> + * that can accommodate the size we want to map, and that is properly
>>> + * aligned to facilitate huge page table entries matching actual
>>> + * huge pages or huge page aligned memory in buffer objects. Buffer
>>> objects
>>> + * are assumed to start at huge page boundary pfns (io memory) or be
>>> + * populated by huge pages aligned to the start of the buffer object
>>> + * (system- or coherent memory). Adapted from shmem_get_unmapped_area.
>>> + *
>>> + * Return: aligned user-space address.
>>> + */
>>> +unsigned long drm_get_unmapped_area(struct file *file,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long uaddr, unsigned long len,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long pgoff, unsigned long flags,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â struct drm_vma_offset_manager *mgr)
>>> +{
>>> +Â Â Â unsigned long addr;
>>> +Â Â Â unsigned long inflated_addr;
>>> +Â Â Â struct drm_vma_offset_node *node;
>>> +
>>> +Â Â Â if (len > TASK_SIZE)
>>> +Â Â Â Â Â Â Â return -ENOMEM;
>>> +
>>> +Â Â Â /* Adjust mapping offset to be zero at bo start */
>>> +Â Â Â drm_vma_offset_lock_lookup(mgr);
>>> +Â Â Â node = drm_vma_offset_lookup_locked(mgr, pgoff, 1);
>>> +Â Â Â if (node)
>>> +Â Â Â Â Â Â Â pgoff -= node->vm_node.start;
>>> +Â Â Â drm_vma_offset_unlock_lookup(mgr);
>>> +
>>> +Â Â Â addr = current->mm->get_unmapped_area(file, uaddr, len, pgoff,
>>> flags);
>>> +Â Â Â if (IS_ERR_VALUE(addr))
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if (addr & ~PAGE_MASK)
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if (addr > TASK_SIZE - len)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â if (len < HPAGE_PMD_SIZE)
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â if (flags & MAP_FIXED)
>>> +Â Â Â Â Â Â Â return addr;
>>> +Â Â Â /*
>>> +Â Â Â Â * Our priority is to support MAP_SHARED mapped hugely;
>>> +Â Â Â Â * and support MAP_PRIVATE mapped hugely too, until it is COWed.
>>> +Â Â Â Â * But if caller specified an address hint, respect that as
>>> before.
>>> +Â Â Â Â */
>>> +Â Â Â if (uaddr)
>>> +Â Â Â Â Â Â Â return addr;
>>> +
>>> +Â Â Â inflated_addr = drm_addr_inflate(addr, len, pgoff, flags,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â HPAGE_PMD_SIZE);
>>> +
>>> +Â Â Â if (IS_ENABLED(CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD) &&
>>> +Â Â Â Â Â Â Â len >= HPAGE_PUD_SIZE)
>>> +Â Â Â Â Â Â Â inflated_addr = drm_addr_inflate(inflated_addr, len, pgoff,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â flags, HPAGE_PUD_SIZE);
>>> +Â Â Â return inflated_addr;
>>> +}
>>> +#else /* CONFIG_TRANSPARENT_HUGEPAGE */
>>> +unsigned long drm_get_unmapped_area(struct file *file,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long uaddr, unsigned long len,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long pgoff, unsigned long flags,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â struct drm_vma_offset_manager *mgr)
>>> +{
>>> +Â Â Â return current->mm->get_unmapped_area(file, uaddr, len, pgoff,
>>> flags);
>>> +}
>>> +#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
>>> +EXPORT_SYMBOL_GPL(drm_get_unmapped_area);
>>> diff --git a/include/drm/drm_file.h b/include/drm/drm_file.h
>>> index 67af60bb527a..4719cc80d547 100644
>>> --- a/include/drm/drm_file.h
>>> +++ b/include/drm/drm_file.h
>>> @@ -386,5 +386,10 @@ void drm_event_cancel_free(struct drm_device *dev,
>>> Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â struct drm_pending_event *p);
>>> Â void drm_send_event_locked(struct drm_device *dev, struct
>>> drm_pending_event *e);
>>> Â void drm_send_event(struct drm_device *dev, struct
>>> drm_pending_event *e);
>>> +struct drm_vma_offset_manager;
>>> +unsigned long drm_get_unmapped_area(struct file *file,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long uaddr, unsigned long len,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â unsigned long pgoff, unsigned long flags,
>>> +Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â struct drm_vma_offset_manager *mgr);
>>> Â Â #endif /* _DRM_FILE_H_ */
>
>
More information about the dri-devel
mailing list