RFC: multitouch support v2

Alex Elsayed eternaleye at gmail.com
Thu Dec 22 18:54:07 PST 2011


Kristian Høgsberg <krh at ...> writes:
> Right... in the MPX sense, right?  So you could have a keyboard and
> mouse combo controlling one pointer/kb focus and the touch screen
> being its own master device.  Then maybe you could have one person
> using the touch screen UI, and another person using the kb/mouse
> combo.  That's kind of far fetched, of course, but I think the main
> point is that there's no inherent association between a kb/mouse combo
> and a touch screen.  On the other hand, what about a setup with two
> mouse/kb combos (master devices) and a touch screen... you'd expect
> tapping a window on the touch screen to set kb focus, but if you have
> multiple master kbs, which kb focus do you set?  Maybe we're just
> doomed for trying to make both pointer and direct touch interaction
> work in the same UI.

One use case you seem to be forgetting is that there are mouse-type
devices like recent Synaptics touchpads that *also* do multitouch.
Multitouch != touchscreen. One way to solve this might be to make 
touchscreens a pointer device *with no associated keyboard device*,
or at least none attached to actual hardware. In XInput, you can create
a new master pair with a real pointer, but only an XTest keyboard. A 
dummy, if you will.



More information about the wayland-devel mailing list