[PATCH v3] Documentation: gpu: Mention the requirements for new properties

Pekka Paalanen ppaalanen at gmail.com
Fri Jun 18 08:55:38 UTC 2021


On Thu, 17 Jun 2021 16:37:14 +0300
Laurent Pinchart <laurent.pinchart at ideasonboard.com> wrote:

> Hi Pekka,
> 
> On Thu, Jun 17, 2021 at 02:33:11PM +0300, Pekka Paalanen wrote:
> > On Thu, 17 Jun 2021 13:29:48 +0300 Laurent Pinchart wrote:  
> > > On Thu, Jun 17, 2021 at 10:27:01AM +0300, Pekka Paalanen wrote:  
> > > > On Thu, 17 Jun 2021 00:05:24 +0300 Laurent Pinchart wrote:    
> > > > > On Tue, Jun 15, 2021 at 01:16:56PM +0300, Pekka Paalanen wrote:    

...

> > > > > > That is where the documented tolerances come into play.      
> > > > > 
> > > > > This is something I've experimented with a while ago, when developing
> > > > > automated tests for the rcar-du driver. When playing with different
> > > > > input images we had to constantly increases tolerances, up to a point
> > > > > where the tests started to miss real problems :-(    
> > > > 
> > > > What should we infer from that? That the hardware is broken and
> > > > exposing those KMS properties is a false promise?    
> > > 
> > > No, just that the scaler doesn't document the internal hardware
> > > implementation (number of taps in the filters, coefficients, rounding,
> > > ...). That's the rule, not the exception, and it doesn't prevent correct
> > > operation, images get scaled in a reproducible way (the same input
> > > produces the same output).
> > >   
> > > > If a driver on certain hardware cannot correctly implement a KMS
> > > > property over the full domain of the input space, should that driver
> > > > then simply not expose the KMS property at all?    
> > > 
> > > The properties involved here would the the SRC and CRTC rectangles for
> > > the planes. They don't document pixel-perfect scaling :-)
> > >   
> > > > But I would assume that the vendor still wants to expose the features
> > > > in upstream kernels, yet they cannot use the standard KMS properties
> > > > for that. Should the driver then expose vendor-specific properties with
> > > > the disclaimer that the result is not always what one would expect, so
> > > > that userspace written and tested explicitly for that hardware can
> > > > still work?
> > > > 
> > > > That is, a sufficient justification for a vendor-specific KMS property
> > > > would be that a standard property already exists, but the hardware is
> > > > too buggy to make it work. IOW, give up trying to make sense.    
> > > 
> > > It's not just about buggy hardware, it's also about implementation
> > > specificities, such as rounding, filters, order of operations in the
> > > color management pipeline (it's relatively easy when you only have two
> > > LUTs and a CCM matrix, but if you through 3D LUTs and other tonemapping
> > > features in the mix, not all hardware will implement the same pipeline),
> > > or various types of image compression (this device implements a
> > > "near-lossless" compression scheme that reduces the memory bandwidth by
> > > 50% for instance).  
> > 
> > Rounding shouldn't result in needing wide tolerances.
> > 
> > Filters are more difficult, but at least we can factor them out when
> > testing other things. Filters could be tested in isolation with some
> > image difference metrics rather than per-pixel independent comparisons.  
> 
> The metrics I was using had both a tolerance on the pixel value and on
> the number of pixels accepted outside of the value tolerance. I'm sure
> we can improve it (perhaps taking locality into account), but that's
> heuristics, and keeping heuristics working across a wide variety of use
> cases is hard.

Hi Laurent,

I was thinking of using a more, um, scientific error measures, e.g.
sum or squared errors (SSE) or average SSE over the whole
re-scaled/filtered image result, ignoring pixels outside of that.

What one normally uses when matching images in computer vision, for
example. Could even add a threshold such that simple rounding-level
errors would not even register in SSE. I'm sure there is plenty of
literature on that, but it may be behind a paywall like IEEE.

SSE may or may not need to computed from light-linear pixel values,
too. If one wanted to go even further, I'm sure there are computational
models about human color and brightness difference sensitivity that
could be used to weigh the errors.


> The filter I mentioned, by the way, is the scaler filter. Out of
> curiosity, do any of the devices you work on document with pixel-perfect
> precision how the hardware scaler is implemented ?

I don't work on drivers, so wouldn't even look for hardware docs. I go
by what KMS UAPI documents because that is the API I work with and
nothing else. And yes, I ignore all the scaling filter issues for now.

Because I don't have a way to get feedback (writeback connectors maybe
not existing and not hooked up in Weston quite yet), testing
scaling/filtering precision has not been on-topic yet. Right now I'm
interested in color correctness rather than geometrical filtering. For
traditional color management, the expected pixel values are quite
precise.

> > The order of operations in the color management pipeline is very
> > important. We can't work with "whatever". All the variability in
> > hardware is exactly why I have been calling out for defined abstract
> > color pipeline in the DRM UAPI docs, so that drivers would know which
> > properties to map to their elements, so that userspace can have any
> > possibility of using them correctly. If the hardware has a block
> > that doesn't fit in the abstract pipeline, you get to add things to the
> > abstract pipeline, or invent a whole another abstract pipeline and
> > document that.  
> 
> One very typical difference between devices is the order of the
> processing blocks. By modelling the KMS pipeline as degamma -> ccm ->
> gamma, we can accommodate hardware that have any combination of
> [1-2] * 1D LUTs + 1 * CCM. Now, throw one 3D LUT into the mix, at

But you cannot represent pipelines like
1D LUT -> 1D LUT -> CCM
because the abstract pipeline just doesn't have the elements for that.
OTOH, maybe that ordering does not even make sense to have in hardware?
So maybe not all combinations are actually needed.

> different points in the pipeline depending on the device, and it will
> start getting complicated, even if the use case is quite simple and
> common. This is getting a bit out of topic, but how would you solve this
> one in particular ?

By defining all the points in the abstract color pipeline where a 3D
LUT could exist. Then each point would probably need its own KMS
property.

We already have the KMS pipeline exactly as degamma -> ctm -> gamma and
drivers need to respect that order.

If the combinatorial explosion gets out of hand, maybe we need a KMS
property to switch to whole another abstract pipeline which defines a
different ordering on the same and/or different KMS properties.

From what I've learnt recently, if you have a 3D LUT, you want a 1D LUT
on each side of it for memory vs. precision optimization. And after the
degamma -> ctm -> gamma pipeline you may want one more ctm for
RGB-to-YCbCr conversion. So I have hope that the abstract pipeline with
all actually implemented hardware features might not go totally out of
hand.

> > Lossy compression needs its own KMS properties to ensure it can be
> > disabled when necessary. Near-lossless is not lossless, if a difference
> > can be measured. The driver or hardware cannot guess if the end user is
> > ok with "near-lossless" or not, so you have to be conservative and
> > assume not ok, offering an opt-in for lossy.  
> 
> Sure, but what would be the barrier to entry for such a property that
> would enable the compression (it could actually be a pixel format
> modifier) ? Would it only need to be documented ? Would we need a
> software implementation in VKMS and/or in IGT ? The compression
> algorithm is proprietary and not documented, so the latter can't be
> done.

Good questions. Shows that the idea of strictly requiring a VKMS
implementation won't fly, which is what I expected.

Saying it could be a pixel format modifier is a really good point. A
modifier cannot be the only thing to control it. Userspace does not
decode modifiers, so it cannot filter it out when it wants lossless
pixels. There must be something else to control it.

As a userspace dev, I would be ok with documenting a KMS property as
"improves blah blah, but also does significant violence to your
pixels", so I would know that this is something I need to consider.

You could argue that all KMS properties do violence to pixels, but
that's not a useful definition. It would just mean that in some use
cases I would never off-load anything to KMS. Depending on the use case
that might still be true even if the errors were limited to reasonable
rounding errors. I need an idea of how much error does KMS processing
do, and ultimately I expect compositors to also need a last resort
button for "do not trust KMS processing at all" which then makes the
display server always use the exact same simplest possible KMS
configuration and let the end user deal with the KMS errors via
color-profiling his monitor.

It's quite a different thing to have color processing elements in an
unexpected order in the pipeline than it is to have a scaling filter
doing slightly unknown operations.

> > ...
> >   
> > > > My underlying assumption is that generic userspace will not use
> > > > vendor-specific properties.    
> > > 
> > > I expect some amount of device-specific code in userspace, yes.   
> > 
> > If we had a reliable way to test device-specific code without the
> > hardware and automatically in CI, then maybe.
> >   
> > > There are usually large variations in how the hardware exposes access to
> > > a given feature, which leads to code having to convert the standard API
> > > parameters to hardware parameters. To a large extend this can be done in
> > > drivers, but for some more complex features, it may put too much burden
> > > on the kernel. There's a reason mesa is a userspace stack :-)  
> > 
> > If we get a Khronos standardised 2D composition API... oh wait.
> > 
> > Nothing wrong with userspace libraries, but it does mean that driver
> > developers need to contribute to them, like they do to Mesa. Is there
> > any of that going on for KMS?  
> 
> Not that I'm aware of, but I think that's a direction we can consider
> seriously.

That would be awesome if the API is generic.

> > In the mean time, DRM UAPI basically must define a 2D composition API,
> > or the new KMS properties will not see use outside of vendor trees.


Thanks,
pq
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 833 bytes
Desc: OpenPGP digital signature
URL: <https://lists.freedesktop.org/archives/dri-devel/attachments/20210618/d2d36d84/attachment.sig>


More information about the dri-devel mailing list