[Intel-gfx] [PATCH 05/16] drm/i915/vm_bind: Implement bind and unbind of object
Niranjana Vishwanathapura
niranjana.vishwanathapura at intel.com
Thu Sep 29 16:38:22 UTC 2022
On Thu, Sep 29, 2022 at 11:49:30AM +0100, Matthew Auld wrote:
>On 28/09/2022 07:19, Niranjana Vishwanathapura wrote:
>>Add uapi and implement support for bind and unbind of an
>>object at the specified GPU virtual addresses.
>>
>>The vm_bind mode is not supported in legacy execbuf2 ioctl.
>>It will be supported only in the newer execbuf3 ioctl.
>>
>>Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura at intel.com>
>>Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan at intel.com>
>>Signed-off-by: Andi Shyti <andi.shyti at linux.intel.com>
>>---
>> drivers/gpu/drm/i915/Makefile | 1 +
>> .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 5 +
>> drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h | 26 ++
>> .../drm/i915/gem/i915_gem_vm_bind_object.c | 306 ++++++++++++++++++
>> drivers/gpu/drm/i915/gt/intel_gtt.c | 10 +
>> drivers/gpu/drm/i915/gt/intel_gtt.h | 17 +
>> drivers/gpu/drm/i915/i915_driver.c | 3 +
>> drivers/gpu/drm/i915/i915_vma.c | 1 +
>> drivers/gpu/drm/i915/i915_vma_types.h | 14 +
>> include/uapi/drm/i915_drm.h | 112 +++++++
>> 10 files changed, 495 insertions(+)
>> create mode 100644 drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
>> create mode 100644 drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c
>>
>>diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
>>index a26edcdadc21..9bf939ef18ea 100644
>>--- a/drivers/gpu/drm/i915/Makefile
>>+++ b/drivers/gpu/drm/i915/Makefile
>>@@ -166,6 +166,7 @@ gem-y += \
>> gem/i915_gem_ttm_move.o \
>> gem/i915_gem_ttm_pm.o \
>> gem/i915_gem_userptr.o \
>>+ gem/i915_gem_vm_bind_object.o \
>> gem/i915_gem_wait.o \
>> gem/i915_gemfs.o
>> i915-y += \
>>diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>>index cd75b0ca2555..f85f10cf9c34 100644
>>--- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>>+++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
>>@@ -781,6 +781,11 @@ static int eb_select_context(struct i915_execbuffer *eb)
>> if (unlikely(IS_ERR(ctx)))
>> return PTR_ERR(ctx);
>>+ if (ctx->vm->vm_bind_mode) {
>>+ i915_gem_context_put(ctx);
>>+ return -EOPNOTSUPP;
>>+ }
>>+
>> eb->gem_context = ctx;
>> if (i915_gem_context_has_full_ppgtt(ctx))
>> eb->invalid_flags |= EXEC_OBJECT_NEEDS_GTT;
>>diff --git a/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
>>new file mode 100644
>>index 000000000000..36262a6357b5
>>--- /dev/null
>>+++ b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
>>@@ -0,0 +1,26 @@
>>+/* SPDX-License-Identifier: MIT */
>>+/*
>>+ * Copyright © 2022 Intel Corporation
>>+ */
>>+
>>+#ifndef __I915_GEM_VM_BIND_H
>>+#define __I915_GEM_VM_BIND_H
>>+
>>+#include <linux/types.h>
>>+
>>+struct drm_device;
>>+struct drm_file;
>>+struct i915_address_space;
>>+struct i915_vma;
>>+
>>+struct i915_vma *
>>+i915_gem_vm_bind_lookup_vma(struct i915_address_space *vm, u64 va);
>>+
>>+int i915_gem_vm_bind_ioctl(struct drm_device *dev, void *data,
>>+ struct drm_file *file);
>>+int i915_gem_vm_unbind_ioctl(struct drm_device *dev, void *data,
>>+ struct drm_file *file);
>>+
>>+void i915_gem_vm_unbind_all(struct i915_address_space *vm);
>>+
>>+#endif /* __I915_GEM_VM_BIND_H */
>>diff --git a/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c
>>new file mode 100644
>>index 000000000000..e529162abd2c
>>--- /dev/null
>>+++ b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c
>>@@ -0,0 +1,306 @@
>>+// SPDX-License-Identifier: MIT
>>+/*
>>+ * Copyright © 2022 Intel Corporation
>>+ */
>>+
>>+#include <uapi/drm/i915_drm.h>
>>+
>>+#include <linux/interval_tree_generic.h>
>>+
>>+#include "gem/i915_gem_context.h"
>>+#include "gem/i915_gem_vm_bind.h"
>>+
>>+#include "gt/intel_gpu_commands.h"
>>+
>>+#define START(node) ((node)->start)
>>+#define LAST(node) ((node)->last)
>>+
>>+INTERVAL_TREE_DEFINE(struct i915_vma, rb, u64, __subtree_last,
>>+ START, LAST, static inline, i915_vm_bind_it)
>>+
>>+#undef START
>>+#undef LAST
>>+
>>+/**
>>+ * DOC: VM_BIND/UNBIND ioctls
>>+ *
>>+ * DRM_I915_GEM_VM_BIND/UNBIND ioctls allows UMD to bind/unbind GEM buffer
>>+ * objects (BOs) or sections of a BOs at specified GPU virtual addresses on a
>>+ * specified address space (VM). Multiple mappings can map to the same physical
>>+ * pages of an object (aliasing). These mappings (also referred to as persistent
>>+ * mappings) will be persistent across multiple GPU submissions (execbuf calls)
>>+ * issued by the UMD, without user having to provide a list of all required
>>+ * mappings during each submission (as required by older execbuf mode).
>>+ *
>>+ * The VM_BIND/UNBIND calls allow UMDs to request a timeline out fence for
>>+ * signaling the completion of bind/unbind operation.
>>+ *
>>+ * VM_BIND feature is advertised to user via I915_PARAM_VM_BIND_VERSION.
>>+ * User has to opt-in for VM_BIND mode of binding for an address space (VM)
>>+ * during VM creation time via I915_VM_CREATE_FLAGS_USE_VM_BIND extension.
>>+ *
>>+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
>>+ * are not ordered. Furthermore, parts of the VM_BIND/UNBIND operations can be
>>+ * done asynchronously, when valid out fence is specified.
>>+ *
>>+ * VM_BIND locking order is as below.
>>+ *
>>+ * 1) vm_bind_lock mutex will protect vm_bind lists. This lock is taken in
>>+ * vm_bind/vm_unbind ioctl calls, in the execbuf path and while releasing the
>>+ * mapping.
>>+ *
>>+ * In future, when GPU page faults are supported, we can potentially use a
>>+ * rwsem instead, so that multiple page fault handlers can take the read
>>+ * side lock to lookup the mapping and hence can run in parallel.
>>+ * The older execbuf mode of binding do not need this lock.
>>+ *
>>+ * 2) The object's dma-resv lock will protect i915_vma state and needs
>>+ * to be held while binding/unbinding a vma in the async worker and while
>>+ * updating dma-resv fence list of an object. Note that private BOs of a VM
>>+ * will all share a dma-resv object.
>>+ *
>>+ * 3) Spinlock/s to protect some of the VM's lists like the list of
>>+ * invalidated vmas (due to eviction and userptr invalidation) etc.
>>+ */
>>+
>>+/**
>>+ * i915_gem_vm_bind_lookup_vma() - lookup for persistent vma mapped at a
>>+ * specified address
>>+ * @vm: virtual address space to look for persistent vma
>>+ * @va: starting address where vma is mapped
>>+ *
>>+ * Retrieves the persistent vma mapped address @va from the @vm's vma tree.
>>+ *
>>+ * Returns vma pointer on success, NULL on failure.
>>+ */
>>+struct i915_vma *
>>+i915_gem_vm_bind_lookup_vma(struct i915_address_space *vm, u64 va)
>>+{
>>+ lockdep_assert_held(&vm->vm_bind_lock);
>>+
>>+ return i915_vm_bind_it_iter_first(&vm->va, va, va);
>>+}
>>+
>>+static void i915_gem_vm_bind_remove(struct i915_vma *vma, bool release_obj)
>>+{
>>+ lockdep_assert_held(&vma->vm->vm_bind_lock);
>>+
>>+ list_del_init(&vma->vm_bind_link);
>>+ i915_vm_bind_it_remove(vma, &vma->vm->va);
>>+
>>+ /* Release object */
>>+ if (release_obj)
>>+ i915_gem_object_put(vma->obj);
>>+}
>>+
>>+static int i915_gem_vm_unbind_vma(struct i915_address_space *vm,
>>+ struct drm_i915_gem_vm_unbind *va)
>>+{
>>+ struct drm_i915_gem_object *obj;
>>+ struct i915_vma *vma;
>>+ int ret;
>>+
>>+ ret = mutex_lock_interruptible(&vm->vm_bind_lock);
>>+ if (ret)
>>+ return ret;
>>+
>>+ va->start = gen8_noncanonical_addr(va->start);
>>+ vma = i915_gem_vm_bind_lookup_vma(vm, va->start);
>>+
>>+ if (!vma)
>>+ ret = -ENOENT;
>>+ else if (vma->size != va->length)
>>+ ret = -EINVAL;
>>+
>>+ if (ret) {
>>+ mutex_unlock(&vm->vm_bind_lock);
>>+ return ret;
>>+ }
>>+
>>+ i915_gem_vm_bind_remove(vma, false);
>>+
>>+ mutex_unlock(&vm->vm_bind_lock);
>>+
>>+ /* Destroy vma and then release object */
>>+ obj = vma->obj;
>
>No locks held here when touching vma->obj? Maybe a small comment
>explaining the safety would be good.
>
We do holding the object reference, so we should be good.
It is documented in vm_bind call when we take object reference.
>>+ ret = i915_gem_object_lock(obj, NULL);
>
>lock_interruptible() ?
Actually, we shouldn't be checking for return error here.
We are not taking interruptible lock so that we don't have to
deal with having an error.
>
>>+ if (ret)
>>+ return ret;
>
>IIUC we have already called bind_remove() above, so vm_bind_lookup()
>can't find it any more? Is the user meant to call this again, if we
>can't grab the object lock? Are we not leaking the below object_put()
>until the entire driver fd is nuked? Can we not grab the vm_bind_lock
>and then object_lock, before calling the bind_remove()? And then on
>lock error userspace can just call this again? Sorry if I'm
>misunderstanding something here.
As mentioned above, we don't expect any error here and I will remove
this error check here. That should fix the problem.
Yah, I guess we can take the object lock also before calling bind_remove()
as well, but I think it is not needed.
>
>>+
>>+ i915_vma_destroy(vma);
>>+ i915_gem_object_unlock(obj);
>>+
>>+ i915_gem_object_put(obj);
>>+
>>+ return 0;
>>+}
>>+
>>+/**
>>+ * i915_gem_vm_unbind_all() - unbind all persistent mappings from an
>>+ * address space
>>+ * @vm: Address spece to remove persistent mappings from
>>+ *
>>+ * Unbind all userspace requested vm_bind mappings from @vm.
>>+ */
>>+void i915_gem_vm_unbind_all(struct i915_address_space *vm)
>>+{
>>+ struct i915_vma *vma, *t;
>>+
>>+ mutex_lock(&vm->vm_bind_lock);
>>+ list_for_each_entry_safe(vma, t, &vm->vm_bind_list, vm_bind_link)
>>+ i915_gem_vm_bind_remove(vma, true);
>>+ list_for_each_entry_safe(vma, t, &vm->vm_bound_list, vm_bind_link)
>>+ i915_gem_vm_bind_remove(vma, true);
>>+ mutex_unlock(&vm->vm_bind_lock);
>>+}
>>+
>>+static struct i915_vma *vm_bind_get_vma(struct i915_address_space *vm,
>>+ struct drm_i915_gem_object *obj,
>>+ struct drm_i915_gem_vm_bind *va)
>>+{
>>+ struct i915_gtt_view view;
>>+ struct i915_vma *vma;
>>+
>>+ va->start = gen8_noncanonical_addr(va->start);
>>+ vma = i915_gem_vm_bind_lookup_vma(vm, va->start);
>>+ if (vma)
>>+ return ERR_PTR(-EEXIST);
>>+
>>+ view.type = I915_GTT_VIEW_PARTIAL;
>>+ view.partial.offset = va->offset >> PAGE_SHIFT;
>>+ view.partial.size = va->length >> PAGE_SHIFT;
>>+ vma = i915_vma_create_persistent(obj, vm, &view);
>>+ if (IS_ERR(vma))
>>+ return vma;
>>+
>>+ vma->start = va->start;
>>+ vma->last = va->start + va->length - 1;
>>+
>>+ return vma;
>>+}
>>+
>>+static int i915_gem_vm_bind_obj(struct i915_address_space *vm,
>>+ struct drm_i915_gem_vm_bind *va,
>>+ struct drm_file *file)
>>+{
>>+ struct drm_i915_gem_object *obj;
>>+ struct i915_vma *vma = NULL;
>>+ struct i915_gem_ww_ctx ww;
>>+ u64 pin_flags;
>>+ int ret = 0;
>>+
>>+ if (!vm->vm_bind_mode)
>>+ return -EOPNOTSUPP;
>
>Also need to check MBZ on the extensions somewhere.
>
Ok, will do.
>>+
>>+ obj = i915_gem_object_lookup(file, va->handle);
>>+ if (!obj)
>>+ return -ENOENT;
>>+
>>+ if (!va->length ||
>>+ !IS_ALIGNED(va->offset | va->length,
>>+ i915_gem_object_max_page_size(obj->mm.placements,
>>+ obj->mm.n_placements)) ||
>>+ range_overflows_t(u64, va->offset, va->length, obj->base.size)) {
>
>Nit: could maybe split this into multiple different checks for better
>readability.
>
Ok, let me see.
>>+ ret = -EINVAL;
>>+ goto put_obj;
>>+ }
>>+
>>+ ret = mutex_lock_interruptible(&vm->vm_bind_lock);
>>+ if (ret)
>>+ goto put_obj;
>>+
>>+ vma = vm_bind_get_vma(vm, obj, va);
>>+ if (IS_ERR(vma)) {
>>+ ret = PTR_ERR(vma);
>>+ goto unlock_vm;
>>+ }
>>+
>>+ pin_flags = va->start | PIN_OFFSET_FIXED | PIN_USER;
>
>va->start for sure needs some sanity checking first. Evil user could
>be setting the low bits here (0-11), which would cause havoc with our
>vma flags.
>
Ok, will do.
Regards,
Niranjana
>>+
>>+ for_i915_gem_ww(&ww, ret, true) {
>>+ ret = i915_gem_object_lock(vma->obj, &ww);
>>+ if (ret)
>>+ continue;
>>+
>>+ ret = i915_vma_pin_ww(vma, &ww, 0, 0, pin_flags);
>>+ if (ret)
>>+ continue;
>>+
>>+ /* Make it evictable */
>>+ __i915_vma_unpin(vma);
>>+
>>+ list_add_tail(&vma->vm_bind_link, &vm->vm_bound_list);
>>+ i915_vm_bind_it_insert(vma, &vm->va);
>>+
>>+ /* Hold object reference until vm_unbind */
>>+ i915_gem_object_get(vma->obj);
>>+ }
>>+
>>+ if (ret)
>>+ i915_vma_destroy(vma);
>>+unlock_vm:
>>+ mutex_unlock(&vm->vm_bind_lock);
>>+put_obj:
>>+ i915_gem_object_put(obj);
>>+
>>+ return ret;
>>+}
>>+
>>+/**
>>+ * i915_gem_vm_bind_ioctl() - ioctl function for binding a section of object
>>+ * at a specified virtual address
>>+ * @dev: drm_device pointer
>>+ * @data: ioctl data structure
>>+ * @file: drm_file pointer
>>+ *
>>+ * Adds the specified persistent mapping (virtual address to a section of an
>>+ * object) and binds it in the device page table.
>>+ *
>>+ * Returns 0 on success, error code on failure.
>>+ */
>>+int i915_gem_vm_bind_ioctl(struct drm_device *dev, void *data,
>>+ struct drm_file *file)
>>+{
>>+ struct drm_i915_gem_vm_bind *args = data;
>>+ struct i915_address_space *vm;
>>+ int ret;
>>+
>>+ vm = i915_gem_vm_lookup(file->driver_priv, args->vm_id);
>>+ if (unlikely(!vm))
>>+ return -ENOENT;
>>+
>>+ ret = i915_gem_vm_bind_obj(vm, args, file);
>>+
>>+ i915_vm_put(vm);
>>+ return ret;
>>+}
>>+
>>+/**
>>+ * i915_gem_vm_unbind_ioctl() - ioctl function for unbinding a mapping at a
>>+ * specified virtual address
>>+ * @dev: drm_device pointer
>>+ * @data: ioctl data structure
>>+ * @file: drm_file pointer
>>+ *
>>+ * Removes the persistent mapping at the specified address and unbinds it
>>+ * from the device page table.
>>+ *
>>+ * Returns 0 on success, error code on failure. -ENOENT is returned if the
>>+ * specified mapping is not found.
>>+ */
>>+int i915_gem_vm_unbind_ioctl(struct drm_device *dev, void *data,
>>+ struct drm_file *file)
>>+{
>>+ struct drm_i915_gem_vm_unbind *args = data;
>>+ struct i915_address_space *vm;
>>+ int ret;
>>+
>>+ vm = i915_gem_vm_lookup(file->driver_priv, args->vm_id);
>>+ if (unlikely(!vm))
>>+ return -ENOENT;
>>+
>>+ ret = i915_gem_vm_unbind_vma(vm, args);
>>+
>>+ i915_vm_put(vm);
>>+ return ret;
>>+}
>>diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.c b/drivers/gpu/drm/i915/gt/intel_gtt.c
>>index b67831833c9a..0daa70c6ed0d 100644
>>--- a/drivers/gpu/drm/i915/gt/intel_gtt.c
>>+++ b/drivers/gpu/drm/i915/gt/intel_gtt.c
>>@@ -12,6 +12,7 @@
>> #include "gem/i915_gem_internal.h"
>> #include "gem/i915_gem_lmem.h"
>>+#include "gem/i915_gem_vm_bind.h"
>> #include "i915_trace.h"
>> #include "i915_utils.h"
>> #include "intel_gt.h"
>>@@ -176,6 +177,8 @@ int i915_vm_lock_objects(struct i915_address_space *vm,
>> void i915_address_space_fini(struct i915_address_space *vm)
>> {
>> drm_mm_takedown(&vm->mm);
>>+ GEM_BUG_ON(!RB_EMPTY_ROOT(&vm->va.rb_root));
>>+ mutex_destroy(&vm->vm_bind_lock);
>> }
>> /**
>>@@ -202,6 +205,8 @@ static void __i915_vm_release(struct work_struct *work)
>> struct i915_address_space *vm =
>> container_of(work, struct i915_address_space, release_work);
>>+ i915_gem_vm_unbind_all(vm);
>>+
>> __i915_vm_close(vm);
>> /* Synchronize async unbinds. */
>>@@ -282,6 +287,11 @@ void i915_address_space_init(struct i915_address_space *vm, int subclass)
>> INIT_LIST_HEAD(&vm->bound_list);
>> INIT_LIST_HEAD(&vm->unbound_list);
>>+
>>+ vm->va = RB_ROOT_CACHED;
>>+ INIT_LIST_HEAD(&vm->vm_bind_list);
>>+ INIT_LIST_HEAD(&vm->vm_bound_list);
>>+ mutex_init(&vm->vm_bind_lock);
>> }
>> void *__px_vaddr(struct drm_i915_gem_object *p)
>>diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.h b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>index c0ca53cba9f0..b52061858161 100644
>>--- a/drivers/gpu/drm/i915/gt/intel_gtt.h
>>+++ b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>@@ -259,6 +259,23 @@ struct i915_address_space {
>> */
>> struct list_head unbound_list;
>>+ /**
>>+ * @vm_bind_mode: flag to indicate vm_bind method of binding
>>+ *
>>+ * True: allow only vm_bind method of binding.
>>+ * False: allow only legacy execbuff method of binding.
>>+ */
>>+ bool vm_bind_mode:1;
>>+
>>+ /** @vm_bind_lock: Mutex to protect @vm_bind_list and @vm_bound_list */
>>+ struct mutex vm_bind_lock;
>>+ /** @vm_bind_list: List of vm_binding in process */
>>+ struct list_head vm_bind_list;
>>+ /** @vm_bound_list: List of vm_binding completed */
>>+ struct list_head vm_bound_list;
>>+ /* @va: tree of persistent vmas */
>>+ struct rb_root_cached va;
>>+
>> /* Global GTT */
>> bool is_ggtt:1;
>>diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c
>>index fb3826dabe8b..c3a9a5031cdb 100644
>>--- a/drivers/gpu/drm/i915/i915_driver.c
>>+++ b/drivers/gpu/drm/i915/i915_driver.c
>>@@ -69,6 +69,7 @@
>> #include "gem/i915_gem_ioctls.h"
>> #include "gem/i915_gem_mman.h"
>> #include "gem/i915_gem_pm.h"
>>+#include "gem/i915_gem_vm_bind.h"
>> #include "gt/intel_gt.h"
>> #include "gt/intel_gt_pm.h"
>> #include "gt/intel_rc6.h"
>>@@ -1893,6 +1894,8 @@ static const struct drm_ioctl_desc i915_ioctls[] = {
>> DRM_IOCTL_DEF_DRV(I915_QUERY, i915_query_ioctl, DRM_RENDER_ALLOW),
>> DRM_IOCTL_DEF_DRV(I915_GEM_VM_CREATE, i915_gem_vm_create_ioctl, DRM_RENDER_ALLOW),
>> DRM_IOCTL_DEF_DRV(I915_GEM_VM_DESTROY, i915_gem_vm_destroy_ioctl, DRM_RENDER_ALLOW),
>>+ DRM_IOCTL_DEF_DRV(I915_GEM_VM_BIND, i915_gem_vm_bind_ioctl, DRM_RENDER_ALLOW),
>>+ DRM_IOCTL_DEF_DRV(I915_GEM_VM_UNBIND, i915_gem_vm_unbind_ioctl, DRM_RENDER_ALLOW),
>> };
>> /*
>>diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
>>index 5839e1f55f00..33f910473263 100644
>>--- a/drivers/gpu/drm/i915/i915_vma.c
>>+++ b/drivers/gpu/drm/i915/i915_vma.c
>>@@ -239,6 +239,7 @@ vma_create(struct drm_i915_gem_object *obj,
>> spin_unlock(&obj->vma.lock);
>> mutex_unlock(&vm->mutex);
>>+ INIT_LIST_HEAD(&vma->vm_bind_link);
>> return vma;
>> err_unlock:
>>diff --git a/drivers/gpu/drm/i915/i915_vma_types.h b/drivers/gpu/drm/i915/i915_vma_types.h
>>index 2200f1f103ba..f56ac07c6cfa 100644
>>--- a/drivers/gpu/drm/i915/i915_vma_types.h
>>+++ b/drivers/gpu/drm/i915/i915_vma_types.h
>>@@ -296,6 +296,20 @@ struct i915_vma {
>> /** This object's place on the active/inactive lists */
>> struct list_head vm_link;
>>+ /** @vm_bind_link: node for the vm_bind related lists of vm */
>>+ struct list_head vm_bind_link;
>>+
>>+ /** Interval tree structures for persistent vma */
>>+
>>+ /** @rb: node for the interval tree of vm for persistent vmas */
>>+ struct rb_node rb;
>>+ /** @start: start endpoint of the rb node */
>>+ u64 start;
>>+ /** @last: Last endpoint of the rb node */
>>+ u64 last;
>>+ /** @__subtree_last: last in subtree */
>>+ u64 __subtree_last;
>>+
>> struct list_head obj_link; /* Link in the object's VMA list */
>> struct rb_node obj_node;
>> struct hlist_node obj_hash;
>>diff --git a/include/uapi/drm/i915_drm.h b/include/uapi/drm/i915_drm.h
>>index 520ad2691a99..9760564b4693 100644
>>--- a/include/uapi/drm/i915_drm.h
>>+++ b/include/uapi/drm/i915_drm.h
>>@@ -470,6 +470,8 @@ typedef struct _drm_i915_sarea {
>> #define DRM_I915_GEM_VM_CREATE 0x3a
>> #define DRM_I915_GEM_VM_DESTROY 0x3b
>> #define DRM_I915_GEM_CREATE_EXT 0x3c
>>+#define DRM_I915_GEM_VM_BIND 0x3d
>>+#define DRM_I915_GEM_VM_UNBIND 0x3e
>> /* Must be kept compact -- no holes */
>> #define DRM_IOCTL_I915_INIT DRM_IOW( DRM_COMMAND_BASE + DRM_I915_INIT, drm_i915_init_t)
>>@@ -534,6 +536,8 @@ typedef struct _drm_i915_sarea {
>> #define DRM_IOCTL_I915_QUERY DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_QUERY, struct drm_i915_query)
>> #define DRM_IOCTL_I915_GEM_VM_CREATE DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_CREATE, struct drm_i915_gem_vm_control)
>> #define DRM_IOCTL_I915_GEM_VM_DESTROY DRM_IOW (DRM_COMMAND_BASE + DRM_I915_GEM_VM_DESTROY, struct drm_i915_gem_vm_control)
>>+#define DRM_IOCTL_I915_GEM_VM_BIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_BIND, struct drm_i915_gem_vm_bind)
>>+#define DRM_IOCTL_I915_GEM_VM_UNBIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_UNBIND, struct drm_i915_gem_vm_unbind)
>> /* Allow drivers to submit batchbuffers directly to hardware, relying
>> * on the security mechanisms provided by hardware.
>>@@ -3717,6 +3721,114 @@ struct drm_i915_gem_create_ext_protected_content {
>> /* ID of the protected content session managed by i915 when PXP is active */
>> #define I915_PROTECTED_CONTENT_DEFAULT_SESSION 0xf
>>+/**
>>+ * struct drm_i915_gem_vm_bind - VA to object mapping to bind.
>>+ *
>>+ * This structure is passed to VM_BIND ioctl and specifies the mapping of GPU
>>+ * virtual address (VA) range to the section of an object that should be bound
>>+ * in the device page table of the specified address space (VM).
>>+ * The VA range specified must be unique (ie., not currently bound) and can
>>+ * be mapped to whole object or a section of the object (partial binding).
>>+ * Multiple VA mappings can be created to the same section of the object
>>+ * (aliasing).
>>+ *
>>+ * The @start, @offset and @length must be 4K page aligned. However the DG2
>>+ * and XEHPSDV has 64K page size for device local memory and has compact page
>>+ * table. On those platforms, for binding device local-memory objects, the
>>+ * @start, @offset and @length must be 64K aligned. Also, UMDs should not mix
>>+ * the local memory 64K page and the system memory 4K page bindings in the same
>>+ * 2M range.
>>+ *
>>+ * Error code -EINVAL will be returned if @start, @offset and @length are not
>>+ * properly aligned. In version 1 (See I915_PARAM_VM_BIND_VERSION), error code
>>+ * -ENOSPC will be returned if the VA range specified can't be reserved.
>>+ *
>>+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
>>+ * are not ordered. Furthermore, parts of the VM_BIND operation can be done
>>+ * asynchronously, if valid @fence is specified.
>>+ */
>>+struct drm_i915_gem_vm_bind {
>>+ /** @vm_id: VM (address space) id to bind */
>>+ __u32 vm_id;
>>+
>>+ /** @handle: Object handle */
>>+ __u32 handle;
>>+
>>+ /** @start: Virtual Address start to bind */
>>+ __u64 start;
>>+
>>+ /** @offset: Offset in object to bind */
>>+ __u64 offset;
>>+
>>+ /** @length: Length of mapping to bind */
>>+ __u64 length;
>>+
>>+ /**
>>+ * @flags: Currently reserved, MBZ.
>>+ *
>>+ * Note that @fence carries its own flags.
>>+ */
>>+ __u64 flags;
>>+
>>+ /** @rsvd: Reserved, MBZ */
>>+ __u64 rsvd[2];
>>+
>>+ /**
>>+ * @extensions: Zero-terminated chain of extensions.
>>+ *
>>+ * For future extensions. See struct i915_user_extension.
>>+ */
>>+ __u64 extensions;
>>+};
>>+
>>+/**
>>+ * struct drm_i915_gem_vm_unbind - VA to object mapping to unbind.
>>+ *
>>+ * This structure is passed to VM_UNBIND ioctl and specifies the GPU virtual
>>+ * address (VA) range that should be unbound from the device page table of the
>>+ * specified address space (VM). VM_UNBIND will force unbind the specified
>>+ * range from device page table without waiting for any GPU job to complete.
>>+ * It is UMDs responsibility to ensure the mapping is no longer in use before
>>+ * calling VM_UNBIND.
>>+ *
>>+ * If the specified mapping is not found, the ioctl will simply return without
>>+ * any error.
>>+ *
>>+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
>>+ * are not ordered. Furthermore, parts of the VM_UNBIND operation can be done
>>+ * asynchronously, if valid @fence is specified.
>>+ */
>>+struct drm_i915_gem_vm_unbind {
>>+ /** @vm_id: VM (address space) id to bind */
>>+ __u32 vm_id;
>>+
>>+ /** @rsvd: Reserved, MBZ */
>>+ __u32 rsvd;
>>+
>>+ /** @start: Virtual Address start to unbind */
>>+ __u64 start;
>>+
>>+ /** @length: Length of mapping to unbind */
>>+ __u64 length;
>>+
>>+ /**
>>+ * @flags: Currently reserved, MBZ.
>>+ *
>>+ * Note that @fence carries its own flags.
>>+ */
>>+ __u64 flags;
>>+
>>+ /** @rsvd2: Reserved, MBZ */
>>+ __u64 rsvd2[2];
>>+
>>+ /**
>>+ * @extensions: Zero-terminated chain of extensions.
>>+ *
>>+ * For future extensions. See struct i915_user_extension.
>>+ */
>>+ __u64 extensions;
>>+};
>>+
>> #if defined(__cplusplus)
>> }
>> #endif
More information about the Intel-gfx
mailing list