[Intel-gfx] [PATCH v6 3/3] drm/doc/rfc: VM_BIND uapi definition

Niranjana Vishwanathapura niranjana.vishwanathapura at intel.com
Thu Jun 30 16:22:18 UTC 2022


On Thu, Jun 30, 2022 at 08:59:09AM +0100, Tvrtko Ursulin wrote:
>
>On 30/06/2022 07:08, Niranjana Vishwanathapura wrote:
>>On Wed, Jun 29, 2022 at 05:33:49PM -0700, Zanoni, Paulo R wrote:
>>>On Sat, 2022-06-25 at 18:49 -0700, Niranjana Vishwanathapura wrote:
>>>>VM_BIND and related uapi definitions
>>>>
>>>>v2: Reduce the scope to simple Mesa use case.
>>>>v3: Expand VM_UNBIND documentation and add
>>>>    I915_GEM_VM_BIND/UNBIND_FENCE_VALID
>>>>    and I915_GEM_VM_BIND_TLB_FLUSH flags.
>>>>v4: Remove I915_GEM_VM_BIND_TLB_FLUSH flag and add additional
>>>>    documentation for vm_bind/unbind.
>>>>v5: Remove TLB flush requirement on VM_UNBIND.
>>>>    Add version support to stage implementation.
>>>>v6: Define and use drm_i915_gem_timeline_fence structure for
>>>>    all timeline fences.
>>>>v7: Rename I915_PARAM_HAS_VM_BIND to I915_PARAM_VM_BIND_VERSION.
>>>>    Update documentation on async vm_bind/unbind and versioning.
>>>>    Remove redundant vm_bind/unbind FENCE_VALID flag, execbuf3
>>>>    batch_count field and I915_EXEC3_SECURE flag.
>>>>
>>>>Signed-off-by: Niranjana Vishwanathapura 
>>>><niranjana.vishwanathapura at intel.com>
>>>>Reviewed-by: Daniel Vetter <daniel.vetter at ffwll.ch>
>>>>---
>>>> Documentation/gpu/rfc/i915_vm_bind.h | 280 +++++++++++++++++++++++++++
>>>> 1 file changed, 280 insertions(+)
>>>> create mode 100644 Documentation/gpu/rfc/i915_vm_bind.h
>>>>
>>>>diff --git a/Documentation/gpu/rfc/i915_vm_bind.h 
>>>>b/Documentation/gpu/rfc/i915_vm_bind.h
>>>>new file mode 100644
>>>>index 000000000000..a93e08bceee6
>>>>--- /dev/null
>>>>+++ b/Documentation/gpu/rfc/i915_vm_bind.h
>>>>@@ -0,0 +1,280 @@
>>>>+/* SPDX-License-Identifier: MIT */
>>>>+/*
>>>>+ * Copyright © 2022 Intel Corporation
>>>>+ */
>>>>+
>>>>+/**
>>>>+ * DOC: I915_PARAM_VM_BIND_VERSION
>>>>+ *
>>>>+ * VM_BIND feature version supported.
>>>>+ * See typedef drm_i915_getparam_t param.
>>>>+ *
>>>>+ * Specifies the VM_BIND feature version supported.
>>>>+ * The following versions of VM_BIND have been defined:
>>>>+ *
>>>>+ * 0: No VM_BIND support.
>>>>+ *
>>>>+ * 1: In VM_UNBIND calls, the UMD must specify the exact 
>>>>mappings created
>>>>+ *    previously with VM_BIND, the ioctl will not support 
>>>>unbinding multiple
>>>>+ *    mappings or splitting them. Similarly, VM_BIND calls will 
>>>>not replace
>>>>+ *    any existing mappings.
>>>>+ *
>>>>+ * 2: The restrictions on unbinding partial or multiple mappings is
>>>>+ *    lifted, Similarly, binding will replace any mappings in 
>>>>the given range.
>>>>+ *
>>>>+ * See struct drm_i915_gem_vm_bind and struct drm_i915_gem_vm_unbind.
>>>>+ */
>>>>+#define I915_PARAM_VM_BIND_VERSION   57
>>>>+
>>>>+/**
>>>>+ * DOC: I915_VM_CREATE_FLAGS_USE_VM_BIND
>>>>+ *
>>>>+ * Flag to opt-in for VM_BIND mode of binding during VM creation.
>>>>+ * See struct drm_i915_gem_vm_control flags.
>>>>+ *
>>>>+ * The older execbuf2 ioctl will not support VM_BIND mode of operation.
>>>>+ * For VM_BIND mode, we have new execbuf3 ioctl which will not 
>>>>accept any
>>>>+ * execlist (See struct drm_i915_gem_execbuffer3 for more details).
>>>>+ */
>>>>+#define I915_VM_CREATE_FLAGS_USE_VM_BIND     (1 << 0)
>>>>+
>>>>+/* VM_BIND related ioctls */
>>>>+#define DRM_I915_GEM_VM_BIND         0x3d
>>>>+#define DRM_I915_GEM_VM_UNBIND               0x3e
>>>>+#define DRM_I915_GEM_EXECBUFFER3     0x3f
>>>>+
>>>>+#define DRM_IOCTL_I915_GEM_VM_BIND           
>>>>DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_BIND, struct 
>>>>drm_i915_gem_vm_bind)
>>>>+#define DRM_IOCTL_I915_GEM_VM_UNBIND         
>>>>DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_UNBIND, struct 
>>>>drm_i915_gem_vm_bind)
>>>>+#define DRM_IOCTL_I915_GEM_EXECBUFFER3               
>>>>DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_EXECBUFFER3, struct 
>>>>drm_i915_gem_execbuffer3)
>>>>+
>>>>+/**
>>>>+ * struct drm_i915_gem_timeline_fence - An input or output 
>>>>timeline fence.
>>>>+ *
>>>>+ * The operation will wait for input fence to signal.
>>>>+ *
>>>>+ * The returned output fence will be signaled after the 
>>>>completion of the
>>>>+ * operation.
>>>>+ */
>>>>+struct drm_i915_gem_timeline_fence {
>>>>+     /** @handle: User's handle for a drm_syncobj to wait on or 
>>>>signal. */
>>>>+     __u32 handle;
>>>>+
>>>>+     /**
>>>>+      * @flags: Supported flags are:
>>>>+      *
>>>>+      * I915_TIMELINE_FENCE_WAIT:
>>>>+      * Wait for the input fence before the operation.
>>>>+      *
>>>>+      * I915_TIMELINE_FENCE_SIGNAL:
>>>>+      * Return operation completion fence as output.
>>>>+      */
>>>>+     __u32 flags;
>>>>+#define I915_TIMELINE_FENCE_WAIT            (1 << 0)
>>>>+#define I915_TIMELINE_FENCE_SIGNAL          (1 << 1)
>>>>+#define __I915_TIMELINE_FENCE_UNKNOWN_FLAGS 
>>>>(-(I915_TIMELINE_FENCE_SIGNAL << 1))
>>>>+
>>>>+     /**
>>>>+      * @value: A point in the timeline.
>>>>+      * Value must be 0 for a binary drm_syncobj. A Value of 0 for a
>>>>+      * timeline drm_syncobj is invalid as it turns a 
>>>>drm_syncobj into a
>>>>+      * binary one.
>>>>+      */
>>>>+     __u64 value;
>>>>+};
>>>>+
>>>>+/**
>>>>+ * struct drm_i915_gem_vm_bind - VA to object mapping to bind.
>>>>+ *
>>>>+ * This structure is passed to VM_BIND ioctl and specifies the 
>>>>mapping of GPU
>>>>+ * virtual address (VA) range to the section of an object that 
>>>>should be bound
>>>>+ * in the device page table of the specified address space (VM).
>>>>+ * The VA range specified must be unique (ie., not currently 
>>>>bound) and can
>>>>+ * be mapped to whole object or a section of the object 
>>>>(partial binding).
>>>>+ * Multiple VA mappings can be created to the same section of 
>>>>the object
>>>>+ * (aliasing).
>>>>+ *
>>>>+ * The @start, @offset and @length must be 4K page aligned. 
>>>>However the DG2
>>>>+ * and XEHPSDV has 64K page size for device local-memory and 
>>>>has compact page
>>>>+ * table. On those platforms, for binding device local-memory 
>>>>objects, the
>>>>+ * @start must be 2M aligned, @offset and @length must be 64K aligned.
>>>>+ * Also, for such mappings, i915 will reserve the whole 2M 
>>>>range for it so as
>>>>+ * to not allow multiple mappings in that 2M range (Compact 
>>>>page tables do not
>>>>+ * allow 64K page and 4K page bindings in the same 2M range).
>>>>+ *
>>>>+ * Error code -EINVAL will be returned if @start, @offset and 
>>>>@length are not
>>>>+ * properly aligned. In version 1 (See 
>>>>I915_PARAM_VM_BIND_VERSION), error code
>>>>+ * -ENOSPC will be returned if the VA range specified can't be 
>>>>reserved.
>>>>+ *
>>>>+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads 
>>>>concurrently
>>>>+ * are not ordered. Furthermore, parts of the VM_BIND operation 
>>>>can be done
>>>>+ * asynchronously, if valid @fence is specified.
>>>
>>>Does that mean that if I don't provide @fence, then this ioctl will be
>>>synchronous (i.e., when it returns, the memory will be guaranteed to be
>>>bound)? The text is kinda implying that, but from one of your earlier
>>>replies to Tvrtko, that doesn't seem to be the case. I guess we could
>>>change the text to make this more explicit.
>>>
>>
>>Yes, I thought, if user doesn't specify the out fence, KMD better make
>>the ioctl synchronous by waiting until the binding finishes before
>>returning. Otherwise, UMD has no way to ensure binding is complete and
>>UMD must pass in out fence for VM_BIND calls.
>
>This problematic angle is exactly what I raised and I did not 
>understand you were suggesting sync behaviour back then.
>
>I suggested a possible execbuf3 extension which makes it wait for any 
>pending (un)bind activity on a VM. Sounds better to me than making 
>everything sync for the use case of N binds followed by 1 execbuf. 
>*If* userspace wants an easy "fire and forget" mode for such use case, 
>rather than having to use a fence on all.
>

This is a good optimization. But it creates some synchronization between
VM_BIND and execbuf3. Based on discussion in IRC, looks like folks are
Ok with waiting in VM_BIND call if out fence is not specified by UMD.
So, we can go with that for now.

Niranjana

>Regards,
>
>Tvrtko
>
>>But latest comment form Daniel on other thread might suggest 
>>something else.
>>Daniel, can you comment?
>>
>>>In addition, previously we had the guarantee that an execbuf ioctl
>>>would wait for all the pending vm_bind operations to finish before
>>>doing anything. Do we still have this guarantee or do we have to make
>>>use of the fences now?
>>>
>>
>>No, we don't have that anymore (execbuf is decoupled from VM_BIND).
>>Execbuf3 submission will not wait for any previous VM_BIND to finish.
>>UMD must pass in VM_BIND out fence as in fence for execbuf3 to ensure
>>that.
>>
>>>>+ */
>>>>+struct drm_i915_gem_vm_bind {
>>>>+     /** @vm_id: VM (address space) id to bind */
>>>>+     __u32 vm_id;
>>>>+
>>>>+     /** @handle: Object handle */
>>>>+     __u32 handle;
>>>>+
>>>>+     /** @start: Virtual Address start to bind */
>>>>+     __u64 start;
>>>>+
>>>>+     /** @offset: Offset in object to bind */
>>>>+     __u64 offset;
>>>>+
>>>>+     /** @length: Length of mapping to bind */
>>>>+     __u64 length;
>>>>+
>>>>+     /**
>>>>+      * @flags: Supported flags are:
>>>>+      *
>>>>+      * I915_GEM_VM_BIND_READONLY:
>>>>+      * Mapping is read-only.
>>>
>>>Can you please explain what happens when we try to write to a range
>>>that's bound as read-only?
>>>
>>
>>It will be mapped as read-only in device page table. Hence any
>>write access will fail. I would expect a CAT error reported.
>>
>>I am seeing that currently the page table R/W setting is based
>>on whether BO is readonly or not (UMDs can request a userptr
>>BO to readonly). We can make this READONLY here as a subset.
>>ie., if BO is readonly, the mappings must be readonly. If BO
>>is not readonly, then the mapping can be either readonly or
>>not.
>>
>>But if Mesa doesn't have a use for this, then we can remove
>>this flag for now.
>>
>>>
>>>>+      *
>>>>+      * I915_GEM_VM_BIND_CAPTURE:
>>>>+      * Capture this mapping in the dump upon GPU error.
>>>>+      */
>>>>+     __u64 flags;
>>>>+#define I915_GEM_VM_BIND_READONLY    (1 << 1)
>>>>+#define I915_GEM_VM_BIND_CAPTURE     (1 << 2)
>>>>+
>>>>+     /**
>>>>+      * @fence: Timeline fence for bind completion signaling.
>>>>+      *
>>>>+      * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
>>>>+      * is invalid, and an error will be returned.
>>>>+      */
>>>>+     struct drm_i915_gem_timeline_fence fence;
>>>>+
>>>>+     /**
>>>>+      * @extensions: Zero-terminated chain of extensions.
>>>>+      *
>>>>+      * For future extensions. See struct i915_user_extension.
>>>>+      */
>>>>+     __u64 extensions;
>>>>+};
>>>>+
>>>>+/**
>>>>+ * struct drm_i915_gem_vm_unbind - VA to object mapping to unbind.
>>>>+ *
>>>>+ * This structure is passed to VM_UNBIND ioctl and specifies 
>>>>the GPU virtual
>>>>+ * address (VA) range that should be unbound from the device 
>>>>page table of the
>>>>+ * specified address space (VM). VM_UNBIND will force unbind 
>>>>the specified
>>>>+ * range from device page table without waiting for any GPU job 
>>>>to complete.
>>>>+ * It is UMDs responsibility to ensure the mapping is no longer 
>>>>in use before
>>>>+ * calling VM_UNBIND.
>>>>+ *
>>>>+ * If the specified mapping is not found, the ioctl will simply 
>>>>return without
>>>>+ * any error.
>>>>+ *
>>>>+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads 
>>>>concurrently
>>>>+ * are not ordered. Furthermore, parts of the VM_UNBIND 
>>>>operation can be done
>>>>+ * asynchronously, if valid @fence is specified.
>>>>+ */
>>>>+struct drm_i915_gem_vm_unbind {
>>>>+     /** @vm_id: VM (address space) id to bind */
>>>>+     __u32 vm_id;
>>>>+
>>>>+     /** @rsvd: Reserved, MBZ */
>>>>+     __u32 rsvd;
>>>>+
>>>>+     /** @start: Virtual Address start to unbind */
>>>>+     __u64 start;
>>>>+
>>>>+     /** @length: Length of mapping to unbind */
>>>>+     __u64 length;
>>>>+
>>>>+     /** @flags: Currently reserved, MBZ */
>>>>+     __u64 flags;
>>>>+
>>>>+     /**
>>>>+      * @fence: Timeline fence for unbind completion signaling.
>>>>+      *
>>>>+      * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
>>>>+      * is invalid, and an error will be returned.
>>>>+      */
>>>>+     struct drm_i915_gem_timeline_fence fence;
>>>>+
>>>>+     /**
>>>>+      * @extensions: Zero-terminated chain of extensions.
>>>>+      *
>>>>+      * For future extensions. See struct i915_user_extension.
>>>>+      */
>>>>+     __u64 extensions;
>>>>+};
>>>>+
>>>>+/**
>>>>+ * struct drm_i915_gem_execbuffer3 - Structure for 
>>>>DRM_I915_GEM_EXECBUFFER3
>>>>+ * ioctl.
>>>>+ *
>>>>+ * DRM_I915_GEM_EXECBUFFER3 ioctl only works in VM_BIND mode 
>>>>and VM_BIND mode
>>>>+ * only works with this ioctl for submission.
>>>>+ * See I915_VM_CREATE_FLAGS_USE_VM_BIND.
>>>>+ */
>>>>+struct drm_i915_gem_execbuffer3 {
>>>>+     /**
>>>>+      * @ctx_id: Context id
>>>>+      *
>>>>+      * Only contexts with user engine map are allowed.
>>>>+      */
>>>>+     __u32 ctx_id;
>>>>+
>>>>+     /**
>>>>+      * @engine_idx: Engine index
>>>>+      *
>>>>+      * An index in the user engine map of the context 
>>>>specified by @ctx_id.
>>>>+      */
>>>>+     __u32 engine_idx;
>>>>+
>>>>+     /**
>>>>+      * @batch_address: Batch gpu virtual address/es.
>>>>+      *
>>>>+      * For normal submission, it is the gpu virtual address of 
>>>>the batch
>>>>+      * buffer. For parallel submission, it is a pointer to an array of
>>>>+      * batch buffer gpu virtual addresses with array size equal to the
>>>>+      * number of (parallel) engines involved in that submission (See
>>>>+      * struct i915_context_engines_parallel_submit).
>>>>+      */
>>>>+     __u64 batch_address;
>>>>+
>>>>+     /** @flags: Currently reserved, MBZ */
>>>>+     __u64 flags;
>>>>+
>>>>+     /** @rsvd1: Reserved, MBZ */
>>>>+     __u32 rsvd1;
>>>>+
>>>>+     /** @fence_count: Number of fences in @timeline_fences array. */
>>>>+     __u32 fence_count;
>>>>+
>>>>+     /**
>>>>+      * @timeline_fences: Pointer to an array of timeline fences.
>>>>+      *
>>>>+      * Timeline fences are of format struct 
>>>>drm_i915_gem_timeline_fence.
>>>>+      */
>>>>+     __u64 timeline_fences;
>>>>+
>>>>+     /** @rsvd2: Reserved, MBZ */
>>>>+     __u64 rsvd2;
>>>>+
>>>
>>>Just out of curiosity: if we can extend behavior with @extensions and
>>>even @flags, why would we need a rsvd2? Perhaps we could kill rsvd2?
>>>
>>
>>True. I added it just in case some requests came up that would require
>>some additional fields. During this review process itself there were
>>some requests. Adding directly here should have a slight performance
>>edge over adding it as an extension (one less copy_from_user).
>>
>>But if folks think this is an overkill, I will remove it.
>>
>>Niranjana
>>
>>>>+     /**
>>>>+      * @extensions: Zero-terminated chain of extensions.
>>>>+      *
>>>>+      * For future extensions. See struct i915_user_extension.
>>>>+      */
>>>>+     __u64 extensions;
>>>>+};
>>>>+
>>>>+/**
>>>>+ * struct drm_i915_gem_create_ext_vm_private - Extension to 
>>>>make the object
>>>>+ * private to the specified VM.
>>>>+ *
>>>>+ * See struct drm_i915_gem_create_ext.
>>>>+ */
>>>>+struct drm_i915_gem_create_ext_vm_private {
>>>>+#define I915_GEM_CREATE_EXT_VM_PRIVATE               2
>>>>+     /** @base: Extension link. See struct i915_user_extension. */
>>>>+     struct i915_user_extension base;
>>>>+
>>>>+     /** @vm_id: Id of the VM to which the object is private */
>>>>+     __u32 vm_id;
>>>>+};
>>>


More information about the Intel-gfx mailing list