[Intel-gfx] [PATCH v3 4/4] drm/doc/rfc: i915 DG1 uAPI
Jason Ekstrand
jason at jlekstrand.net
Tue Apr 20 17:00:51 UTC 2021
On Tue, Apr 20, 2021 at 11:34 AM Tvrtko Ursulin
<tvrtko.ursulin at linux.intel.com> wrote:
>
>
> On 19/04/2021 16:19, Jason Ekstrand wrote:
> > On Mon, Apr 19, 2021 at 7:02 AM Matthew Auld <matthew.auld at intel.com> wrote:
> >>
> >> On 16/04/2021 17:38, Jason Ekstrand wrote:
> >>> On Thu, Apr 15, 2021 at 11:04 AM Matthew Auld <matthew.auld at intel.com> wrote:
> >>>>
> >>>> Add an entry for the new uAPI needed for DG1.
> >>>>
> >>>> v2(Daniel):
> >>>> - include the overall upstreaming plan
> >>>> - add a note for mmap, there are differences here for TTM vs i915
> >>>> - bunch of other suggestions from Daniel
> >>>> v3:
> >>>> (Daniel)
> >>>> - add a note for set/get caching stuff
> >>>> - add some more docs for existing query and extensions stuff
> >>>> - add an actual code example for regions query
> >>>> - bunch of other stuff
> >>>> (Jason)
> >>>> - uAPI change(!):
> >>>> - try a simpler design with the placements extension
> >>>> - rather than have a generic setparam which can cover multiple
> >>>> use cases, have each extension be responsible for one thing
> >>>> only
> >>>>
> >>>> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> >>>> Cc: Joonas Lahtinen <joonas.lahtinen at linux.intel.com>
> >>>> Cc: Jordan Justen <jordan.l.justen at intel.com>
> >>>> Cc: Daniel Vetter <daniel.vetter at intel.com>
> >>>> Cc: Kenneth Graunke <kenneth at whitecape.org>
> >>>> Cc: Jason Ekstrand <jason at jlekstrand.net>
> >>>> Cc: Dave Airlie <airlied at gmail.com>
> >>>> Cc: dri-devel at lists.freedesktop.org
> >>>> Cc: mesa-dev at lists.freedesktop.org
> >>>> ---
> >>>> Documentation/gpu/rfc/i915_gem_lmem.h | 255 ++++++++++++++++++++++++
> >>>> Documentation/gpu/rfc/i915_gem_lmem.rst | 139 +++++++++++++
> >>>> Documentation/gpu/rfc/index.rst | 4 +
> >>>> 3 files changed, 398 insertions(+)
> >>>> create mode 100644 Documentation/gpu/rfc/i915_gem_lmem.h
> >>>> create mode 100644 Documentation/gpu/rfc/i915_gem_lmem.rst
> >>>>
> >>>> diff --git a/Documentation/gpu/rfc/i915_gem_lmem.h b/Documentation/gpu/rfc/i915_gem_lmem.h
> >>>> new file mode 100644
> >>>> index 000000000000..2a82a452e9f2
> >>>> --- /dev/null
> >>>> +++ b/Documentation/gpu/rfc/i915_gem_lmem.h
> >>>> @@ -0,0 +1,255 @@
> >>>> +/*
> >>>> + * Note that drm_i915_query_item and drm_i915_query are existing bits of uAPI.
> >>>> + * For the regions query we are just adding a new query id, so no actual new
> >>>> + * ioctl or anything, but including it here for reference.
> >>>> + */
> >>>> +struct drm_i915_query_item {
> >>>> +#define DRM_I915_QUERY_MEMORY_REGIONS 0xdeadbeaf
> >>>> + ....
> >>>> + __u64 query_id;
> >>>> +
> >>>> + /*
> >>>> + * When set to zero by userspace, this is filled with the size of the
> >>>> + * data to be written at the data_ptr pointer. The kernel sets this
> >>>> + * value to a negative value to signal an error on a particular query
> >>>> + * item.
> >>>> + */
> >>>> + __s32 length;
> >>>> +
> >>>> + __u32 flags;
> >>>> + /*
> >>>> + * Data will be written at the location pointed by data_ptr when the
> >>>> + * value of length matches the length of the data to be written by the
> >>>> + * kernel.
> >>>> + */
> >>>> + __u64 data_ptr;
> >>>> +};
> >>>> +
> >>>> +struct drm_i915_query {
> >>>> + __u32 num_items;
> >>>> + /*
> >>>> + * Unused for now. Must be cleared to zero.
> >>>> + */
> >>>> + __u32 flags;
> >>>> + /*
> >>>> + * This points to an array of num_items drm_i915_query_item structures.
> >>>> + */
> >>>> + __u64 items_ptr;
> >>>> +};
> >>>> +
> >>>> +#define DRM_IOCTL_I915_QUERY DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_QUERY, struct drm_i915_query)
> >>>> +
> >>>> +/**
> >>>> + * enum drm_i915_gem_memory_class
> >>>> + */
> >>>> +enum drm_i915_gem_memory_class {
> >>>> + /** @I915_MEMORY_CLASS_SYSTEM: system memory */
> >>>> + I915_MEMORY_CLASS_SYSTEM = 0,
> >>>> + /** @I915_MEMORY_CLASS_DEVICE: device local-memory */
> >>>> + I915_MEMORY_CLASS_DEVICE,
> >>>> +};
> >>>> +
> >>>> +/**
> >>>> + * struct drm_i915_gem_memory_class_instance
> >>>> + */
> >>>> +struct drm_i915_gem_memory_class_instance {
> >>>> + /** @memory_class: see enum drm_i915_gem_memory_class */
> >>>> + __u16 memory_class;
> >>>> +
> >>>> + /** @memory_instance: which instance */
> >>>> + __u16 memory_instance;
> >>>> +};
> >>>> +
> >>>> +/**
> >>>> + * struct drm_i915_memory_region_info
> >>>> + *
> >>>> + * Describes one region as known to the driver.
> >>>> + *
> >>>> + * Note that we reserve quite a lot of stuff here for potential future work. As
> >>>> + * an example we might want expose the capabilities(see caps) for a given
> >>>> + * region, which could include things like if the region is CPU
> >>>> + * mappable/accessible etc.
> >>>
> >>> I get caps but I'm seriously at a loss as to what the rest of this
> >>> would be used for. Why are caps and flags both there and separate?
> >>> Flags are typically something you set, not query. Also, what's with
> >>> rsvd1 at the end? This smells of substantial over-building to me.
> >>>
> >>> I thought to myself, "maybe I'm missing a future use-case" so I looked
> >>> at the internal tree and none of this is being used there either.
> >>> This indicates to me that either I'm missing something and there's
> >>> code somewhere I don't know about or, with three years of building on
> >>> internal branches, we still haven't proven that any of this is needed.
> >>> If it's the latter, which I strongly suspect, maybe we should drop the
> >>> unnecessary bits and only add them back in if and when we have proof
> >>> that they're useful.
> >>
> >> Do you mean just drop caps/flags here, but keep/inflate rsvd0/rsvd1,
> >> which is less opinionated about future unknowns? If so, makes sense to me.
> >
> > I meant drop flags and rsvd1. We need rsvd0 for padding and I can
> > see some value to caps. We may want to advertise, for instance, what
> > mapping coherency types are available per-heap. But I don't see any
> > use for any of the other fields.
>
> I'd suggest making sure at least enough rsvd fields remain so that flags
> could be added later if needed. Experience from engine info shows that
> both were required in order to extend the query via re-purposing the
> rsvds and adding flag bits to indicate when a certain rsvd contains a
> new piece of information.
Looking at DII, all I see is we started using caps. I already said
I'm fine with caps. I can already imagine some useful ones like
specifying what kinds of mappings we can do.
If we're concerned about more complicated stuff, I argue that we have
no ability to predict what that will look like and so just throwing in
a bunch of __u32 rsvd[N] is blind guessing. I'm seeing a lot of that
in the recently added APIs such as the flags and rsvd[4] in
i915_user_extension. What's that there for? Why can't you put that
information in the extension struct which derives from it? Maybe it's
so that we can extend it. But we added that struct as part of an
extension mechanism!?!
If we want to make things extensible, Vulkan actually provides some
prior art for this in the form of allowing queries to be extended just
like input structs. We could add a __u64 extensions field to
memory_region_info and, if we ever need to query more info, the client
can provide a chain of additional per-region queries. Yeah, there are
problems with it and it gets a bit clunky but it does work pretty
well.
> I probably cannot go into too much detail
> here, but anyway the point is just to make sure too much is not stripped
> out so that instead of simply adding fields/flags we have to add a new
> query in the future. IMO some rsvd fields are not really harmful and if
> they can make things easier in the future why not.
Maybe it's my tired and generally grumpy state of mind but I'm not
particularly favorable towards "why not?" as a justification for
immutable kernel APIs. We've already found a few places where
Zoidberg API design has caused us problems. We need an answer to
"why?" Future extensibility is a potentially valid answer but we need
to do a better job of thinking through it than we have in the past.
--Jason
More information about the Intel-gfx
mailing list