[RFC] Plane color pipeline KMS uAPI
jadahl at redhat.com
Mon May 8 10:03:59 UTC 2023
On Mon, May 08, 2023 at 09:14:18AM +1000, Dave Airlie wrote:
> On Sat, 6 May 2023 at 08:21, Sebastian Wick <sebastian.wick at redhat.com> wrote:
> > On Fri, May 5, 2023 at 10:40 PM Dave Airlie <airlied at gmail.com> wrote:
> > >
> > > On Fri, 5 May 2023 at 01:23, Simon Ser <contact at emersion.fr> wrote:
> > > >
> > > > Hi all,
> > > >
> > > > The goal of this RFC is to expose a generic KMS uAPI to configure the color
> > > > pipeline before blending, ie. after a pixel is tapped from a plane's
> > > > framebuffer and before it's blended with other planes. With this new uAPI we
> > > > aim to reduce the battery life impact of color management and HDR on mobile
> > > > devices, to improve performance and to decrease latency by skipping
> > > > composition on the 3D engine. This proposal is the result of discussions at
> > > > the Red Hat HDR hackfest  which took place a few days ago. Engineers
> > > > familiar with the AMD, Intel and NVIDIA hardware have participated in the
> > > > discussion.
> > > >
> > > > This proposal takes a prescriptive approach instead of a descriptive approach.
> > > > Drivers describe the available hardware blocks in terms of low-level
> > > > mathematical operations, then user-space configures each block. We decided
> > > > against a descriptive approach where user-space would provide a high-level
> > > > description of the colorspace and other parameters: we want to give more
> > > > control and flexibility to user-space, e.g. to be able to replicate exactly the
> > > > color pipeline with shaders and switch between shaders and KMS pipelines
> > > > seamlessly, and to avoid forcing user-space into a particular color management
> > > > policy.
> > >
> > > I'm not 100% sold on the prescriptive here, let's see if someone can
> > > get me over the line with some questions later.
> > >
> > > My feeling is color pipeline hw is not a done deal, and that hw
> > > vendors will be revising/evolving/churning the hw blocks for a while
> > > longer, as there is no real standards in the area to aim for, all the
> > > vendors are mostly just doing whatever gets Windows over the line and
> > > keeps hw engineers happy. So I have some concerns here around forwards
> > > compatibility and hence the API design.
> > >
> > > I guess my main concern is if you expose a bunch of hw blocks and
> > > someone comes up with a novel new thing, will all existing userspace
> > > work, without falling back to shaders?
> > > Do we have minimum guarantees on what hardware blocks have to be
> > > exposed to build a useable pipeline?
> > > If a hardware block goes away in a new silicon revision, do I have to
> > > rewrite my compositor? or will it be expected that the kernel will
> > > emulate the old pipelines on top of whatever new fancy thing exists.
> > I think there are two answers to those questions.
> These aren't selling me much better :-)
> > The first one is that right now KMS already doesn't guarantee that
> > every property is supported on all hardware. The guarantee we have is
> > that properties that are supported on a piece of hardware on a
> > specific kernel will be supported on the same hardware on later
> > kernels. The color pipeline is no different here. For a specific piece
> > of hardware a newer kernel might only change the pipelines in a
> > backwards compatible way and add new pipelines.
> > So to answer your question: if some hardware with a novel pipeline
> > will show up it might not be supported and that's fine. We already
> > have cases where some hardware does not support the gamma lut property
> > but only the CSC property and that breaks night light because we never
> > bothered to write a shader fallback. KMS provides ways to offload work
> > but a generic user space always has to provide a fallback and this
> > doesn't change. Hardware specific user space on the other hand will
> > keep working with the forward compatibility guarantees we want to
> > provide.
> In my mind we've screwed up already, isn't a case to be made for
> continue down the same path.
> The kernel is meant to be a hardware abstraction layer, not just a
> hardware exposure layer. The kernel shouldn't set policy and there are
> cases where it can't act as an abstraction layer (like where you need
> a compiler), but I'm not sold that this case is one of those yet. I'm
> open to being educated here on why it would be.
It would still be an abstraction of the hardware, just that the level
of abstraction is a bit "lower" than your intuition currently tells you
we should have. IMO it's not too different from the kernel providing low
level input events describing what what the hardware can do and does,
with a rather massive user space library (libinput) turning all of that
low level nonsense to actual useful abstractions.
In this case it's the other way around, the kernel provides vendor
independent knobs that describe what the output hardware can do, and
exactly how it does it, and a userspace library turns that into a
different and perhaps more useful abstraction.
I realize input and output is dramatically different, just making a
point about the level of abstraction that is ideal is not necessary "the
more the better".
> > The second answer is that we want to provide a user space library
> > which takes a description of a color pipeline and tries to map that to
> > the available KMS color pipelines. If there is a novel color
> > operation, adding support in this library would then make it possible
> > to offload compatible color pipelines on this new hardware for all
> > consumers of the library. Obviously there is no guarantee that
> > whatever color pipeline compositors come up with can actually be
> > realized on specific hardware but that's just an inherent hardware
> > issue.
> Why does this library need to be in userspace though? If there's a
> library making device dependent decisions, why can't we just make
> those device dependent decisions in the kernel?
Compositors will want to switch between using the KMS color pipeline and
using shaders without visible differences at any point in time, or
predictable visible differences. Lets say for example you are playing a
video, and everything bypasses the GPU, and we're saving important power
and all that. Suddenly the user moves the mouse or touches the screen
and we'll suddenly have overlays that make it necessary to stop
bypassing the GPU and start compositing.
These transitions should not result in any visible difference, and that
is hard/impossible to do perfectly if the level of abstraction is too
high, as implementation details of the pipeline would be hidden. The
decisions the kernel had to make to turn the descriptive declaration
into actual hardware configuration wouldn't be predictable or known.
Userspace needs to know how the kernel implements a pipeline, so that it
can decide if it's "good enough" or perhaps even adapt its compositing
to match it so that it can implement non-glitchy offloading. The
compositor should make the decision if pixel perfect transitions is
mandatory or not, it shouldn't be a policy implemented inside the
It is also within scope of a library that provides the descriptive API
to also know how to handle the fallback, e.g. by providing shaders that
compositors can use when compositing. The kernel is the wrong place to
> This feels like we are trying to go down the Android HWC road, but we
> aren't in that business.
> My thoughts would be userspace has to have some way to describe what
> it wants anyways, otherwise it does sound like I need to update
> mutter, kwin, surfaceflinger, chromeos, gamescope, every time a new HW
> device comes out that operates slightly different to previously
> generations. This isn't the kernel doing hw abstraction at all, it's
> the kernel just opting out of designing interfaces and it isn't
> something I'm sold on.
It is true that for a new generation of hardware that changes the color
pipeline in a way that makes existing userspace no longer able to
off-load compositing, needs an updated userspace, but the grand long
term idea is that one wouldn't update all those compositors, but only
the shared library that provides the descriptive API they all (ideally)
make use of. The compositors still handle interacting with KMS
themselves, but would share this helper library that can help
configuring a subset of the knobs KMS provide to their individual needs.
So indeed the kernel isn't doing all the abstraction, it'd be the kernel
together with a userspace library.
More information about the wayland-devel