client side decorations

Bill Spitzak spitzak at gmail.com
Thu May 12 10:47:06 PDT 2011


microcai wrote:

> They can't care how big a windows is in the pixel, but in the inch.
> 
> People should have different monitors with different DPI. Windows should
> stay same size regardless the DPI.
> 
> Force DPI==96 on every monitor is a stupid idea, and we should avoid it
> on the protocol side.

The reason this had to be done was due to the incredibly stupid idea 
that only *fonts* are measured in points, and every other graphic is 
measured in pixels. This strange idea was on both X and Windows, likely 
due to the initial programs being terminal emulators where there was no 
graphics other than text. What this really means is that there are two 
different coordinate systems for all the graphics, and programmers just 
assumed these two systems always lined up exactly like they did on their 
screen.

After a lot of awful looking output on screens with different DPI, both 
Windows and then X resorted to just forcing the DPI to 96, thus making 
the systems obey the programmer's assumptions. Bad DPI settings are 
still a bug on X, producing ridiculous large and tiny font sizes 
unexpectedly, and this is NEVER wanted.

The correct solution would have been to specify all coordinates in the 
same units, likely 1 unit in the CTM. For practical reasons on 
current-day screens this wants to be a pixel by default, but there is no 
reason a program can't read the DPI and set the CTM to draw actual-size 
graphics.

> I suggest, DPI should also be windows specific, so that compositor can
> *scale it*.

If I understand it right, a Wayland window has both a rectangle measured 
in screen coordinates, and a source image that can be a different size. 
The compositor is expected to transform the source image (scale it) to 
fit in the rectangle.


More information about the wayland-devel mailing list