[Intel-gfx] [PATCH 36/39] drm/i915: Use reservation_object to coordinate userptr get_pages()
Daniel Vetter
daniel at ffwll.ch
Wed Aug 7 20:49:58 UTC 2019
On Fri, Jun 14, 2019 at 08:10:20AM +0100, Chris Wilson wrote:
> Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
Not sure this works, 2 thoughts way down ...
> ---
> drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c | 27 +--
> .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 3 -
> drivers/gpu/drm/i915/gem/i915_gem_internal.c | 30 +--
> .../gpu/drm/i915/gem/i915_gem_object_types.h | 11 +-
> drivers/gpu/drm/i915/gem/i915_gem_pages.c | 137 ++++++++++-
> drivers/gpu/drm/i915/gem/i915_gem_phys.c | 33 ++-
> drivers/gpu/drm/i915/gem/i915_gem_shmem.c | 30 +--
> drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 25 +-
> drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 221 ++++--------------
> .../drm/i915/gem/selftests/huge_gem_object.c | 17 +-
> .../gpu/drm/i915/gem/selftests/huge_pages.c | 45 ++--
> drivers/gpu/drm/i915/gvt/dmabuf.c | 17 +-
> drivers/gpu/drm/i915/i915_drv.h | 9 +-
> drivers/gpu/drm/i915/i915_gem.c | 5 +-
> drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 14 +-
> 15 files changed, 300 insertions(+), 324 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> index 84992d590da5..a44d6d2ef7ed 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c
> @@ -225,33 +225,24 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev,
> return drm_gem_dmabuf_export(dev, &exp_info);
> }
>
> -static int i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +dmabuf_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> - struct sg_table *pages;
> - unsigned int sg_page_sizes;
> -
> - pages = dma_buf_map_attachment(obj->base.import_attach,
> - DMA_BIDIRECTIONAL);
> - if (IS_ERR(pages))
> - return PTR_ERR(pages);
> -
> - sg_page_sizes = i915_sg_page_sizes(pages->sgl);
> -
> - __i915_gem_object_set_pages(obj, pages, sg_page_sizes);
> -
> - return 0;
> + return dma_buf_map_attachment(ctx->object->base.import_attach,
> + DMA_BIDIRECTIONAL);
> }
>
> -static void i915_gem_object_put_pages_dmabuf(struct drm_i915_gem_object *obj,
> - struct sg_table *pages)
> +static void dmabuf_put_pages(struct drm_i915_gem_object *obj,
> + struct sg_table *pages)
> {
> dma_buf_unmap_attachment(obj->base.import_attach, pages,
> DMA_BIDIRECTIONAL);
> }
>
> static const struct drm_i915_gem_object_ops i915_gem_object_dmabuf_ops = {
> - .get_pages = i915_gem_object_get_pages_dmabuf,
> - .put_pages = i915_gem_object_put_pages_dmabuf,
> + .get_pages = dmabuf_get_pages,
> + .put_pages = dmabuf_put_pages,
> };
>
> struct drm_gem_object *i915_gem_prime_import(struct drm_device *dev,
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> index 44bcb681c168..68faf1a71c97 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> @@ -1784,9 +1784,6 @@ static noinline int eb_relocate_slow(struct i915_execbuffer *eb)
> goto out;
> }
>
> - /* A frequent cause for EAGAIN are currently unavailable client pages */
> - flush_workqueue(eb->i915->mm.userptr_wq);
> -
> err = i915_mutex_lock_interruptible(dev);
> if (err) {
> mutex_lock(&dev->struct_mutex);
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_internal.c b/drivers/gpu/drm/i915/gem/i915_gem_internal.c
> index 0c41e04ab8fa..aa0bd5de313b 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_internal.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_internal.c
> @@ -32,11 +32,14 @@ static void internal_free_pages(struct sg_table *st)
> kfree(st);
> }
>
> -static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +internal_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct drm_i915_private *i915 = to_i915(obj->base.dev);
> - struct sg_table *st;
> struct scatterlist *sg;
> + struct sg_table *st;
> unsigned int sg_page_sizes;
> unsigned int npages;
> int max_order;
> @@ -66,12 +69,12 @@ static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj)
> create_st:
> st = kmalloc(sizeof(*st), GFP_KERNEL);
> if (!st)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> npages = obj->base.size / PAGE_SIZE;
> if (sg_alloc_table(st, npages, GFP_KERNEL)) {
> kfree(st);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> sg = st->sgl;
> @@ -117,27 +120,26 @@ static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj)
> goto err;
> }
>
> - /* Mark the pages as dontneed whilst they are still pinned. As soon
> + /*
> + * Mark the pages as dontneed whilst they are still pinned. As soon
> * as they are unpinned they are allowed to be reaped by the shrinker,
> * and the caller is expected to repopulate - the contents of this
> * object are only valid whilst active and pinned.
> */
> obj->mm.madv = I915_MADV_DONTNEED;
>
> - __i915_gem_object_set_pages(obj, st, sg_page_sizes);
> -
> - return 0;
> + *sizes = sg_page_sizes;
> + return st;
>
> err:
> sg_set_page(sg, NULL, 0, 0);
> sg_mark_end(sg);
> internal_free_pages(st);
> -
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> -static void i915_gem_object_put_pages_internal(struct drm_i915_gem_object *obj,
> - struct sg_table *pages)
> +static void internal_put_pages(struct drm_i915_gem_object *obj,
> + struct sg_table *pages)
> {
> i915_gem_gtt_finish_pages(obj, pages);
> internal_free_pages(pages);
> @@ -149,8 +151,8 @@ static void i915_gem_object_put_pages_internal(struct drm_i915_gem_object *obj,
> static const struct drm_i915_gem_object_ops i915_gem_object_internal_ops = {
> .flags = I915_GEM_OBJECT_HAS_STRUCT_PAGE |
> I915_GEM_OBJECT_IS_SHRINKABLE,
> - .get_pages = i915_gem_object_get_pages_internal,
> - .put_pages = i915_gem_object_put_pages_internal,
> + .get_pages = internal_get_pages,
> + .put_pages = internal_put_pages,
> };
>
> /**
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h
> index f792953b8a71..0ea404cfbc1c 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h
> @@ -19,6 +19,7 @@
>
> struct drm_i915_gem_object;
> struct intel_fronbuffer;
> +struct task_struct;
>
> /*
> * struct i915_lut_handle tracks the fast lookups from handle to vma used
> @@ -32,6 +33,11 @@ struct i915_lut_handle {
> u32 handle;
> };
>
> +struct i915_gem_object_get_pages_context {
> + struct drm_i915_gem_object *object;
> + struct task_struct *task;
> +};
> +
> struct drm_i915_gem_object_ops {
> unsigned int flags;
> #define I915_GEM_OBJECT_HAS_STRUCT_PAGE BIT(0)
> @@ -52,9 +58,11 @@ struct drm_i915_gem_object_ops {
> * being released or under memory pressure (where we attempt to
> * reap pages for the shrinker).
> */
> - int (*get_pages)(struct drm_i915_gem_object *obj);
> + struct sg_table *(*get_pages)(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes);
> void (*put_pages)(struct drm_i915_gem_object *obj,
> struct sg_table *pages);
> +
> void (*truncate)(struct drm_i915_gem_object *obj);
> void (*writeback)(struct drm_i915_gem_object *obj);
>
> @@ -252,7 +260,6 @@ struct drm_i915_gem_object {
>
> struct i915_mm_struct *mm;
> struct i915_mmu_object *mmu_object;
> - struct work_struct *work;
> } userptr;
>
> unsigned long scratch;
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> index 6bec301cee79..f65a983248c6 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> @@ -8,16 +8,49 @@
> #include "i915_gem_object.h"
> #include "i915_scatterlist.h"
>
> -void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
> - struct sg_table *pages,
> - unsigned int sg_page_sizes)
> +static DEFINE_SPINLOCK(fence_lock);
> +
> +struct get_pages_work {
> + struct dma_fence dma; /* Must be first for dma_fence_free() */
> + struct i915_sw_fence wait;
> + struct work_struct work;
> + struct i915_gem_object_get_pages_context ctx;
> +};
> +
> +static const char *get_pages_work_driver_name(struct dma_fence *fence)
> +{
> + return DRIVER_NAME;
> +}
> +
> +static const char *get_pages_work_timeline_name(struct dma_fence *fence)
> +{
> + return "allocation";
> +}
> +
> +static void get_pages_work_release(struct dma_fence *fence)
> +{
> + struct get_pages_work *w = container_of(fence, typeof(*w), dma);
> +
> + i915_sw_fence_fini(&w->wait);
> +
> + BUILD_BUG_ON(offsetof(typeof(*w), dma));
> + dma_fence_free(&w->dma);
> +}
> +
> +static const struct dma_fence_ops get_pages_work_ops = {
> + .get_driver_name = get_pages_work_driver_name,
> + .get_timeline_name = get_pages_work_timeline_name,
> + .release = get_pages_work_release,
> +};
> +
> +static void __set_pages(struct drm_i915_gem_object *obj,
> + struct sg_table *pages,
> + unsigned int sg_page_sizes)
> {
> struct drm_i915_private *i915 = to_i915(obj->base.dev);
> unsigned long supported = INTEL_INFO(i915)->page_sizes;
> int i;
>
> - lockdep_assert_held(&obj->mm.lock);
> -
> /* Make the pages coherent with the GPU (flushing any swapin). */
> if (obj->cache_dirty) {
> obj->write_domain = 0;
> @@ -29,8 +62,6 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
> obj->mm.get_page.sg_pos = pages->sgl;
> obj->mm.get_page.sg_idx = 0;
>
> - obj->mm.pages = pages;
> -
> if (i915_gem_object_is_tiled(obj) &&
> i915->quirks & QUIRK_PIN_SWIZZLED_PAGES) {
> GEM_BUG_ON(obj->mm.quirked);
> @@ -38,7 +69,8 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
> obj->mm.quirked = true;
> }
>
> - GEM_BUG_ON(!sg_page_sizes);
> + if (!sg_page_sizes)
> + sg_page_sizes = i915_sg_page_sizes(pages->sgl);
> obj->mm.page_sizes.phys = sg_page_sizes;
>
> /*
> @@ -73,18 +105,105 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
>
> spin_unlock_irqrestore(&i915->mm.obj_lock, flags);
> }
> +}
>
> +static void
> +get_pages_worker(struct work_struct *_work)
> +{
> + struct get_pages_work *work = container_of(_work, typeof(*work), work);
> + struct drm_i915_gem_object *obj = work->ctx.object;
> + struct sg_table *pages;
> + unsigned int sizes = 0;
> +
> + if (!work->dma.error) {
> + pages = obj->ops->get_pages(&work->ctx, &sizes);
> + if (!IS_ERR(pages))
> + __set_pages(obj, pages, sizes);
> + else
> + dma_fence_set_error(&work->dma, PTR_ERR(pages));
> + } else {
> + pages = ERR_PTR(work->dma.error);
> + }
> +
> + obj->mm.pages = pages;
> complete_all(&obj->mm.completion);
> + atomic_dec(&obj->mm.pages_pin_count);
> +
> + i915_gem_object_put(obj);
> + put_task_struct(work->ctx.task);
> +
> + dma_fence_signal(&work->dma);
> + dma_fence_put(&work->dma);
> +}
> +
> +static int __i915_sw_fence_call
> +get_pages_notify(struct i915_sw_fence *fence, enum i915_sw_fence_notify state)
> +{
> + struct get_pages_work *w = container_of(fence, typeof(*w), wait);
> +
> + switch (state) {
> + case FENCE_COMPLETE:
> + if (fence->error)
> + dma_fence_set_error(&w->dma, fence->error);
> + queue_work(system_unbound_wq, &w->work);
> + break;
> +
> + case FENCE_FREE:
> + dma_fence_put(&w->dma);
> + break;
> + }
> +
> + return NOTIFY_DONE;
> }
>
> int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
> {
> + struct get_pages_work *work;
> + int err;
> +
> if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) {
> DRM_DEBUG("Attempting to obtain a purgeable object\n");
> return -EFAULT;
> }
>
> - return obj->ops->get_pages(obj);
> + /* XXX inline? */
> +
> + work = kmalloc(sizeof(*work), GFP_KERNEL);
> + if (!work)
> + return -ENOMEM;
> +
> + dma_fence_init(&work->dma,
> + &get_pages_work_ops,
> + &fence_lock,
> + to_i915(obj->base.dev)->mm.unordered_timeline,
> + 0);
With all the shrinkers (across the tree, not just i915) we have that just
blindly assume you can wait for a fence (at least if you manage to acquire
the reservation_lock) I expect this to deadlock. I think the rule is that
no published fence can depend on any memory allocation anywhere.
Yes it sucks real hard.
> + i915_sw_fence_init(&work->wait, get_pages_notify);
> +
> + work->ctx.object = i915_gem_object_get(obj);
> +
> + work->ctx.task = current;
> + get_task_struct(work->ctx.task);
> +
> + INIT_WORK(&work->work, get_pages_worker);
> +
> + i915_gem_object_lock(obj);
Other bit is nesting reservation_object lock within obj->mm.lock.
obj->mm.lock is a very neat&tidy answer to dealing with bo alloc vs.
shrinker madness. But looking at where dma-buf is headed we'll need to
have the reservation_object nest outside of the obj->mm.lock (because the
importer is going to acquire that for us, before calling down into
exporter code).
Of course on the importer side we're currently nesting the other way
round, and Christian König already set CI on fire figuring that out.
I think what could work is pulling the reservation_object out of
obj->mm.lock, and entirely bypassing it for imported dma-buf (since for
that case dealing with shrinker lolz isn't our problem, but the exporters,
so we don't need the special shrinker lock).
3rd part is just irrational me freaking out when we hide userptr behind a
wait_completion, that didn't work out well last time around because
lockdep fails us. But that's really just an aside.
-Daniel
> + GEM_BUG_ON(!reservation_object_test_signaled_rcu(obj->resv, true));
> + err = i915_sw_fence_await_reservation(&work->wait,
> + obj->resv, NULL,
> + true, I915_FENCE_TIMEOUT,
> + I915_FENCE_GFP);
> + if (err == 0) {
> + reservation_object_add_excl_fence(obj->resv, &work->dma);
> + atomic_inc(&obj->mm.pages_pin_count);
> + } else {
> + dma_fence_set_error(&work->dma, err);
> + }
> + i915_gem_object_unlock(obj);
> +
> + dma_fence_get(&work->dma);
> + i915_sw_fence_commit(&work->wait);
> +
> + return err;
> }
>
> /* Ensure that the associated pages are gathered from the backing storage
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_phys.c b/drivers/gpu/drm/i915/gem/i915_gem_phys.c
> index 2deac933cf59..6b4a5fb52055 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_phys.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_phys.c
> @@ -17,18 +17,21 @@
> #include "i915_gem_object.h"
> #include "i915_scatterlist.h"
>
> -static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +phys_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct address_space *mapping = obj->base.filp->f_mapping;
> struct drm_dma_handle *phys;
> - struct sg_table *st;
> struct scatterlist *sg;
> + struct sg_table *st;
> char *vaddr;
> int i;
> int err;
>
> if (WARN_ON(i915_gem_object_needs_bit17_swizzle(obj)))
> - return -EINVAL;
> + return ERR_PTR(-EINVAL);
>
> /* Always aligning to the object size, allows a single allocation
> * to handle all possible callers, and given typical object sizes,
> @@ -38,7 +41,7 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
> roundup_pow_of_two(obj->base.size),
> roundup_pow_of_two(obj->base.size));
> if (!phys)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> vaddr = phys->vaddr;
> for (i = 0; i < obj->base.size / PAGE_SIZE; i++) {
> @@ -83,19 +86,16 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj)
>
> obj->phys_handle = phys;
>
> - __i915_gem_object_set_pages(obj, st, sg->length);
> -
> - return 0;
> + *sizes = sg->length;
> + return st;
>
> err_phys:
> drm_pci_free(obj->base.dev, phys);
> -
> - return err;
> + return ERR_PTR(err);
> }
>
> static void
> -i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj,
> - struct sg_table *pages)
> +phys_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages)
> {
> __i915_gem_object_release_shmem(obj, pages, false);
>
> @@ -139,8 +139,8 @@ i915_gem_object_release_phys(struct drm_i915_gem_object *obj)
> }
>
> static const struct drm_i915_gem_object_ops i915_gem_phys_ops = {
> - .get_pages = i915_gem_object_get_pages_phys,
> - .put_pages = i915_gem_object_put_pages_phys,
> + .get_pages = phys_get_pages,
> + .put_pages = phys_put_pages,
> .release = i915_gem_object_release_phys,
> };
>
> @@ -193,15 +193,12 @@ int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align)
> if (!IS_ERR_OR_NULL(pages))
> i915_gem_shmem_ops.put_pages(obj, pages);
> mutex_unlock(&obj->mm.lock);
> +
> + wait_for_completion(&obj->mm.completion);
> return 0;
>
> err_xfer:
> obj->ops = &i915_gem_shmem_ops;
> - if (!IS_ERR_OR_NULL(pages)) {
> - unsigned int sg_page_sizes = i915_sg_page_sizes(pages->sgl);
> -
> - __i915_gem_object_set_pages(obj, pages, sg_page_sizes);
> - }
> err_unlock:
> mutex_unlock(&obj->mm.lock);
> return err;
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> index 19d9ecdb2894..c43304e3bada 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> @@ -22,11 +22,13 @@ static void check_release_pagevec(struct pagevec *pvec)
> cond_resched();
> }
>
> -static int shmem_get_pages(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +shmem_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct drm_i915_private *i915 = to_i915(obj->base.dev);
> const unsigned long page_count = obj->base.size / PAGE_SIZE;
> - unsigned long i;
> struct address_space *mapping;
> struct sg_table *st;
> struct scatterlist *sg;
> @@ -37,31 +39,24 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj)
> unsigned int sg_page_sizes;
> struct pagevec pvec;
> gfp_t noreclaim;
> + unsigned long i;
> int ret;
>
> - /*
> - * Assert that the object is not currently in any GPU domain. As it
> - * wasn't in the GTT, there shouldn't be any way it could have been in
> - * a GPU cache
> - */
> - GEM_BUG_ON(obj->read_domains & I915_GEM_GPU_DOMAINS);
> - GEM_BUG_ON(obj->write_domain & I915_GEM_GPU_DOMAINS);
> -
> /*
> * If there's no chance of allocating enough pages for the whole
> * object, bail early.
> */
> if (page_count > totalram_pages())
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> st = kmalloc(sizeof(*st), GFP_KERNEL);
> if (!st)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> rebuild_st:
> if (sg_alloc_table(st, page_count, GFP_KERNEL)) {
> kfree(st);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> /*
> @@ -179,9 +174,8 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj)
> if (i915_gem_object_needs_bit17_swizzle(obj))
> i915_gem_object_do_bit_17_swizzle(obj, st);
>
> - __i915_gem_object_set_pages(obj, st, sg_page_sizes);
> -
> - return 0;
> + *sizes = sg_page_sizes;
> + return st;
>
> err_sg:
> sg_mark_end(sg);
> @@ -209,7 +203,7 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj)
> if (ret == -ENOSPC)
> ret = -ENOMEM;
>
> - return ret;
> + return ERR_PTR(ret);
> }
>
> static void
> @@ -276,8 +270,6 @@ __i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
> struct sg_table *pages,
> bool needs_clflush)
> {
> - GEM_BUG_ON(obj->mm.madv == __I915_MADV_PURGED);
> -
> if (obj->mm.madv == I915_MADV_DONTNEED)
> obj->mm.dirty = false;
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> index 7066044d63cf..28ea06e667cd 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c
> @@ -499,22 +499,19 @@ i915_pages_create_for_stolen(struct drm_device *dev,
> return st;
> }
>
> -static int i915_gem_object_get_pages_stolen(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +stolen_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> - struct sg_table *pages =
> - i915_pages_create_for_stolen(obj->base.dev,
> - obj->stolen->start,
> - obj->stolen->size);
> - if (IS_ERR(pages))
> - return PTR_ERR(pages);
> -
> - __i915_gem_object_set_pages(obj, pages, obj->stolen->size);
> + struct drm_i915_gem_object *obj = ctx->object;
>
> - return 0;
> + return i915_pages_create_for_stolen(obj->base.dev,
> + obj->stolen->start,
> + obj->stolen->size);
> }
>
> -static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj,
> - struct sg_table *pages)
> +static void
> +stolen_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages)
> {
> /* Should only be called from i915_gem_object_release_stolen() */
> sg_free_table(pages);
> @@ -536,8 +533,8 @@ i915_gem_object_release_stolen(struct drm_i915_gem_object *obj)
> }
>
> static const struct drm_i915_gem_object_ops i915_gem_object_stolen_ops = {
> - .get_pages = i915_gem_object_get_pages_stolen,
> - .put_pages = i915_gem_object_put_pages_stolen,
> + .get_pages = stolen_get_pages,
> + .put_pages = stolen_put_pages,
> .release = i915_gem_object_release_stolen,
> };
>
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> index f093deaeb5c0..6748e15bf89a 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> @@ -378,7 +378,7 @@ __i915_mm_struct_free(struct kref *kref)
> mutex_unlock(&mm->i915->mm_lock);
>
> INIT_WORK(&mm->work, __i915_mm_struct_free__worker);
> - queue_work(mm->i915->mm.userptr_wq, &mm->work);
> + schedule_work(&mm->work);
> }
>
> static void
> @@ -393,19 +393,12 @@ i915_gem_userptr_release__mm_struct(struct drm_i915_gem_object *obj)
> obj->userptr.mm = NULL;
> }
>
> -struct get_pages_work {
> - struct work_struct work;
> - struct drm_i915_gem_object *obj;
> - struct task_struct *task;
> -};
> -
> static struct sg_table *
> __i915_gem_userptr_alloc_pages(struct drm_i915_gem_object *obj,
> struct page **pvec, int num_pages)
> {
> unsigned int max_segment = i915_sg_segment_size();
> struct sg_table *st;
> - unsigned int sg_page_sizes;
> int ret;
>
> st = kmalloc(sizeof(*st), GFP_KERNEL);
> @@ -435,131 +428,23 @@ __i915_gem_userptr_alloc_pages(struct drm_i915_gem_object *obj,
> return ERR_PTR(ret);
> }
>
> - sg_page_sizes = i915_sg_page_sizes(st->sgl);
> -
> - __i915_gem_object_set_pages(obj, st, sg_page_sizes);
> -
> return st;
> }
>
> -static void
> -__i915_gem_userptr_get_pages_worker(struct work_struct *_work)
> -{
> - struct get_pages_work *work = container_of(_work, typeof(*work), work);
> - struct drm_i915_gem_object *obj = work->obj;
> - const int npages = obj->base.size >> PAGE_SHIFT;
> - struct page **pvec;
> - int pinned, ret;
> -
> - ret = -ENOMEM;
> - pinned = 0;
> -
> - pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL);
> - if (pvec != NULL) {
> - struct mm_struct *mm = obj->userptr.mm->mm;
> - unsigned int flags = 0;
> -
> - if (!i915_gem_object_is_readonly(obj))
> - flags |= FOLL_WRITE;
> -
> - ret = -EFAULT;
> - if (mmget_not_zero(mm)) {
> - down_read(&mm->mmap_sem);
> - while (pinned < npages) {
> - ret = get_user_pages_remote
> - (work->task, mm,
> - obj->userptr.ptr + pinned * PAGE_SIZE,
> - npages - pinned,
> - flags,
> - pvec + pinned, NULL, NULL);
> - if (ret < 0)
> - break;
> -
> - pinned += ret;
> - }
> - up_read(&mm->mmap_sem);
> - mmput(mm);
> - }
> - }
> -
> - mutex_lock(&obj->mm.lock);
> - if (obj->userptr.work == &work->work) {
> - struct sg_table *pages = ERR_PTR(ret);
> -
> - if (pinned == npages) {
> - pages = __i915_gem_userptr_alloc_pages(obj, pvec,
> - npages);
> - if (!IS_ERR(pages)) {
> - pinned = 0;
> - pages = NULL;
> - }
> - }
> -
> - obj->userptr.work = ERR_CAST(pages);
> - if (IS_ERR(pages))
> - __i915_gem_userptr_set_active(obj, false);
> - }
> - mutex_unlock(&obj->mm.lock);
> -
> - release_pages(pvec, pinned);
> - kvfree(pvec);
> -
> - i915_gem_object_put(obj);
> - put_task_struct(work->task);
> - kfree(work);
> -}
> -
> static struct sg_table *
> -__i915_gem_userptr_get_pages_schedule(struct drm_i915_gem_object *obj)
> -{
> - struct get_pages_work *work;
> -
> - /* Spawn a worker so that we can acquire the
> - * user pages without holding our mutex. Access
> - * to the user pages requires mmap_sem, and we have
> - * a strict lock ordering of mmap_sem, struct_mutex -
> - * we already hold struct_mutex here and so cannot
> - * call gup without encountering a lock inversion.
> - *
> - * Userspace will keep on repeating the operation
> - * (thanks to EAGAIN) until either we hit the fast
> - * path or the worker completes. If the worker is
> - * cancelled or superseded, the task is still run
> - * but the results ignored. (This leads to
> - * complications that we may have a stray object
> - * refcount that we need to be wary of when
> - * checking for existing objects during creation.)
> - * If the worker encounters an error, it reports
> - * that error back to this function through
> - * obj->userptr.work = ERR_PTR.
> - */
> - work = kmalloc(sizeof(*work), GFP_KERNEL);
> - if (work == NULL)
> - return ERR_PTR(-ENOMEM);
> -
> - obj->userptr.work = &work->work;
> -
> - work->obj = i915_gem_object_get(obj);
> -
> - work->task = current;
> - get_task_struct(work->task);
> -
> - INIT_WORK(&work->work, __i915_gem_userptr_get_pages_worker);
> - queue_work(to_i915(obj->base.dev)->mm.userptr_wq, &work->work);
> -
> - return ERR_PTR(-EAGAIN);
> -}
> -
> -static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj)
> +userptr_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> const int num_pages = obj->base.size >> PAGE_SHIFT;
> struct mm_struct *mm = obj->userptr.mm->mm;
> - struct page **pvec;
> struct sg_table *pages;
> - bool active;
> + struct page **pvec;
> int pinned;
> + int ret;
>
> - /* If userspace should engineer that these pages are replaced in
> + /*
> + * If userspace should engineer that these pages are replaced in
> * the vma between us binding this page into the GTT and completion
> * of rendering... Their loss. If they change the mapping of their
> * pages they need to create a new bo to point to the new vma.
> @@ -576,59 +461,58 @@ static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj)
> * egregious cases from causing harm.
> */
>
> - if (obj->userptr.work) {
> - /* active flag should still be held for the pending work */
> - if (IS_ERR(obj->userptr.work))
> - return PTR_ERR(obj->userptr.work);
> - else
> - return -EAGAIN;
> - }
> + pvec = kvmalloc_array(num_pages, sizeof(struct page *),
> + GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
> + if (!pvec)
> + return ERR_PTR(-ENOMEM);
> +
> + __i915_gem_userptr_set_active(obj, true);
>
> - pvec = NULL;
> pinned = 0;
> + ret = -EFAULT;
> + if (mmget_not_zero(mm)) {
> + unsigned int flags;
>
> - if (mm == current->mm) {
> - pvec = kvmalloc_array(num_pages, sizeof(struct page *),
> - GFP_KERNEL |
> - __GFP_NORETRY |
> - __GFP_NOWARN);
> - if (pvec) /* defer to worker if malloc fails */
> - pinned = __get_user_pages_fast(obj->userptr.ptr,
> - num_pages,
> - !i915_gem_object_is_readonly(obj),
> - pvec);
> - }
> + flags = 0;
> + if (!i915_gem_object_is_readonly(obj))
> + flags |= FOLL_WRITE;
>
> - active = false;
> - if (pinned < 0) {
> - pages = ERR_PTR(pinned);
> - pinned = 0;
> - } else if (pinned < num_pages) {
> - pages = __i915_gem_userptr_get_pages_schedule(obj);
> - active = pages == ERR_PTR(-EAGAIN);
> - } else {
> - pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages);
> - active = !IS_ERR(pages);
> + down_read(&mm->mmap_sem);
> + while (pinned < num_pages) {
> + ret = get_user_pages_remote
> + (ctx->task, mm,
> + obj->userptr.ptr + pinned * PAGE_SIZE,
> + num_pages - pinned,
> + flags,
> + pvec + pinned, NULL, NULL);
> + if (ret < 0)
> + break;
> +
> + pinned += ret;
> + }
> + up_read(&mm->mmap_sem);
> + mmput(mm);
> }
> - if (active)
> - __i915_gem_userptr_set_active(obj, true);
>
> - if (IS_ERR(pages))
> + if (ret)
> + pages = ERR_PTR(ret);
> + else
> + pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages);
> + if (IS_ERR(pages)) {
> release_pages(pvec, pinned);
> + __i915_gem_userptr_set_active(obj, false);
> + }
> kvfree(pvec);
>
> - return PTR_ERR_OR_ZERO(pages);
> + return pages;
> }
>
> static void
> -i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj,
> - struct sg_table *pages)
> +userptr_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages)
> {
> struct sgt_iter sgt_iter;
> struct page *page;
>
> - /* Cancel any inflight work and force them to restart their gup */
> - obj->userptr.work = NULL;
> __i915_gem_userptr_set_active(obj, false);
> if (!pages)
> return;
> @@ -669,8 +553,8 @@ static const struct drm_i915_gem_object_ops i915_gem_userptr_ops = {
> .flags = I915_GEM_OBJECT_HAS_STRUCT_PAGE |
> I915_GEM_OBJECT_IS_SHRINKABLE |
> I915_GEM_OBJECT_ASYNC_CANCEL,
> - .get_pages = i915_gem_userptr_get_pages,
> - .put_pages = i915_gem_userptr_put_pages,
> + .get_pages = userptr_get_pages,
> + .put_pages = userptr_put_pages,
> .dmabuf_export = i915_gem_userptr_dmabuf_export,
> .release = i915_gem_userptr_release,
> };
> @@ -786,22 +670,13 @@ i915_gem_userptr_ioctl(struct drm_device *dev,
> return 0;
> }
>
> -int i915_gem_init_userptr(struct drm_i915_private *dev_priv)
> +void i915_gem_init_userptr(struct drm_i915_private *dev_priv)
> {
> mutex_init(&dev_priv->mm_lock);
> hash_init(dev_priv->mm_structs);
> -
> - dev_priv->mm.userptr_wq =
> - alloc_workqueue("i915-userptr-acquire",
> - WQ_HIGHPRI | WQ_UNBOUND,
> - 0);
> - if (!dev_priv->mm.userptr_wq)
> - return -ENOMEM;
> -
> - return 0;
> }
>
> void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv)
> {
> - destroy_workqueue(dev_priv->mm.userptr_wq);
> + mutex_destroy(&dev_priv->mm_lock);
> }
> diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c b/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c
> index 3c5d17b2b670..02e6edce715e 100644
> --- a/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c
> +++ b/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c
> @@ -21,9 +21,12 @@ static void huge_free_pages(struct drm_i915_gem_object *obj,
> kfree(pages);
> }
>
> -static int huge_get_pages(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +huge_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
> + struct drm_i915_gem_object *obj = ctx->object;
> const unsigned long nreal = obj->scratch / PAGE_SIZE;
> const unsigned long npages = obj->base.size / PAGE_SIZE;
> struct scatterlist *sg, *src, *end;
> @@ -32,11 +35,11 @@ static int huge_get_pages(struct drm_i915_gem_object *obj)
>
> pages = kmalloc(sizeof(*pages), GFP);
> if (!pages)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> if (sg_alloc_table(pages, npages, GFP)) {
> kfree(pages);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> sg = pages->sgl;
> @@ -64,14 +67,12 @@ static int huge_get_pages(struct drm_i915_gem_object *obj)
> if (i915_gem_gtt_prepare_pages(obj, pages))
> goto err;
>
> - __i915_gem_object_set_pages(obj, pages, PAGE_SIZE);
> -
> - return 0;
> + *sizes = PAGE_SIZE;
> + return pages;
>
> err:
> huge_free_pages(obj, pages);
> -
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> #undef GFP
> }
>
> diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> index b03a29366bd1..3b93a7337f27 100644
> --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c
> @@ -51,9 +51,12 @@ static void huge_pages_free_pages(struct sg_table *st)
> kfree(st);
> }
>
> -static int get_huge_pages(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +get_huge_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
> + struct drm_i915_gem_object *obj = ctx->object;
> unsigned int page_mask = obj->mm.page_mask;
> struct sg_table *st;
> struct scatterlist *sg;
> @@ -62,11 +65,11 @@ static int get_huge_pages(struct drm_i915_gem_object *obj)
>
> st = kmalloc(sizeof(*st), GFP);
> if (!st)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
> kfree(st);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> rem = obj->base.size;
> @@ -114,16 +117,14 @@ static int get_huge_pages(struct drm_i915_gem_object *obj)
> obj->mm.madv = I915_MADV_DONTNEED;
>
> GEM_BUG_ON(sg_page_sizes != obj->mm.page_mask);
> - __i915_gem_object_set_pages(obj, st, sg_page_sizes);
> -
> - return 0;
> + *sizes = sg_page_sizes;
> + return st;
>
> err:
> sg_set_page(sg, NULL, 0, 0);
> sg_mark_end(sg);
> huge_pages_free_pages(st);
> -
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> static void put_huge_pages(struct drm_i915_gem_object *obj,
> @@ -175,8 +176,11 @@ huge_pages_object(struct drm_i915_private *i915,
> return obj;
> }
>
> -static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +fake_get_huge_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct drm_i915_private *i915 = to_i915(obj->base.dev);
> const u64 max_len = rounddown_pow_of_two(UINT_MAX);
> struct sg_table *st;
> @@ -186,11 +190,11 @@ static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
>
> st = kmalloc(sizeof(*st), GFP);
> if (!st)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) {
> kfree(st);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> /* Use optimal page sized chunks to fill in the sg table */
> @@ -227,13 +231,15 @@ static int fake_get_huge_pages(struct drm_i915_gem_object *obj)
>
> obj->mm.madv = I915_MADV_DONTNEED;
>
> - __i915_gem_object_set_pages(obj, st, sg_page_sizes);
> -
> - return 0;
> + *sizes = sg_page_sizes;
> + return st;
> }
>
> -static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +fake_get_huge_pages_single(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct drm_i915_private *i915 = to_i915(obj->base.dev);
> struct sg_table *st;
> struct scatterlist *sg;
> @@ -241,11 +247,11 @@ static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
>
> st = kmalloc(sizeof(*st), GFP);
> if (!st)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> if (sg_alloc_table(st, 1, GFP)) {
> kfree(st);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> sg = st->sgl;
> @@ -261,9 +267,8 @@ static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj)
>
> obj->mm.madv = I915_MADV_DONTNEED;
>
> - __i915_gem_object_set_pages(obj, st, sg->length);
> -
> - return 0;
> + *sizes = sg->length;
> + return st;
> #undef GFP
> }
>
> diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c
> index 41c8ebc60c63..c2a81939f698 100644
> --- a/drivers/gpu/drm/i915/gvt/dmabuf.c
> +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c
> @@ -36,9 +36,11 @@
>
> #define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12))
>
> -static int vgpu_gem_get_pages(
> - struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +vgpu_gem_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> + struct drm_i915_gem_object *obj = ctx->object;
> struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> struct sg_table *st;
> struct scatterlist *sg;
> @@ -49,17 +51,17 @@ static int vgpu_gem_get_pages(
>
> fb_info = (struct intel_vgpu_fb_info *)obj->gvt_info;
> if (WARN_ON(!fb_info))
> - return -ENODEV;
> + return ERR_PTR(-ENODEV);
>
> st = kmalloc(sizeof(*st), GFP_KERNEL);
> if (unlikely(!st))
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> page_num = obj->base.size >> PAGE_SHIFT;
> ret = sg_alloc_table(st, page_num, GFP_KERNEL);
> if (ret) {
> kfree(st);
> - return ret;
> + return ERR_PTR(ret);
> }
> gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm +
> (fb_info->start >> PAGE_SHIFT);
> @@ -71,9 +73,8 @@ static int vgpu_gem_get_pages(
> sg_dma_len(sg) = PAGE_SIZE;
> }
>
> - __i915_gem_object_set_pages(obj, st, PAGE_SIZE);
> -
> - return 0;
> + *sizes = PAGE_SIZE;
> + return st;
> }
>
> static void vgpu_gem_put_pages(struct drm_i915_gem_object *obj,
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index 9a1ec487e8b1..ef9e4cb49b4a 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -786,13 +786,6 @@ struct i915_gem_mm {
> struct notifier_block vmap_notifier;
> struct shrinker shrinker;
>
> - /**
> - * Workqueue to fault in userptr pages, flushed by the execbuf
> - * when required but otherwise left to userspace to try again
> - * on EAGAIN.
> - */
> - struct workqueue_struct *userptr_wq;
> -
> u64 unordered_timeline;
>
> /* the indicator for dispatch video commands on two BSD rings */
> @@ -2514,7 +2507,7 @@ static inline bool intel_vgpu_active(struct drm_i915_private *dev_priv)
> }
>
> /* i915_gem.c */
> -int i915_gem_init_userptr(struct drm_i915_private *dev_priv);
> +void i915_gem_init_userptr(struct drm_i915_private *dev_priv);
> void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv);
> void i915_gem_sanitize(struct drm_i915_private *i915);
> int i915_gem_init_early(struct drm_i915_private *dev_priv);
> diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
> index 21416b87905e..1571c707ad15 100644
> --- a/drivers/gpu/drm/i915/i915_gem.c
> +++ b/drivers/gpu/drm/i915/i915_gem.c
> @@ -1544,10 +1544,7 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
> dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1);
>
> i915_timelines_init(dev_priv);
> -
> - ret = i915_gem_init_userptr(dev_priv);
> - if (ret)
> - return ret;
> + i915_gem_init_userptr(dev_priv);
>
> ret = intel_uc_init_misc(dev_priv);
> if (ret)
> diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> index 664b7e3e1d7c..dae1f8634a38 100644
> --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
> @@ -54,10 +54,13 @@ static void fake_free_pages(struct drm_i915_gem_object *obj,
> kfree(pages);
> }
>
> -static int fake_get_pages(struct drm_i915_gem_object *obj)
> +static struct sg_table *
> +fake_get_pages(struct i915_gem_object_get_pages_context *ctx,
> + unsigned int *sizes)
> {
> #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
> #define PFN_BIAS 0x1000
> + struct drm_i915_gem_object *obj = ctx->object;
> struct sg_table *pages;
> struct scatterlist *sg;
> unsigned int sg_page_sizes;
> @@ -65,12 +68,12 @@ static int fake_get_pages(struct drm_i915_gem_object *obj)
>
> pages = kmalloc(sizeof(*pages), GFP);
> if (!pages)
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
>
> rem = round_up(obj->base.size, BIT(31)) >> 31;
> if (sg_alloc_table(pages, rem, GFP)) {
> kfree(pages);
> - return -ENOMEM;
> + return ERR_PTR(-ENOMEM);
> }
>
> sg_page_sizes = 0;
> @@ -90,9 +93,8 @@ static int fake_get_pages(struct drm_i915_gem_object *obj)
>
> obj->mm.madv = I915_MADV_DONTNEED;
>
> - __i915_gem_object_set_pages(obj, pages, sg_page_sizes);
> -
> - return 0;
> + *sizes = sg_page_sizes;
> + return pages;
> #undef GFP
> }
>
> --
> 2.20.1
>
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx
--
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
More information about the Intel-gfx
mailing list