[Intel-gfx] [PATCH] drm/i915: Depend on !PREEMPT_RT.
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Wed Mar 2 11:42:35 UTC 2022
On 01/03/2022 15:13, Sebastian Andrzej Siewior wrote:
> On 2022-03-01 14:27:18 [+0000], Tvrtko Ursulin wrote:
>>> you see:
>>> 0003-drm-i915-Use-preempt_disable-enable_rt-where-recomme.patch
>>> 0004-drm-i915-Don-t-disable-interrupts-on-PREEMPT_RT-duri.patch
>>
>> Two for the display folks.
>>
>>> 0005-drm-i915-Don-t-check-for-atomic-context-on-PREEMPT_R.patch
>>
>> What do preempt_disable/enable do on PREEMPT_RT? Thinking if instead the
>> solution could be to always force the !ATOMIC path (for the whole
>> _wait_for_atomic macro) on PREEMPT_RT.
>
> Could be one way to handle it. But please don't disable preemption and
> or interrupts for longer period of time as all of it increases the
> overall latency.
I am looking for your guidance of what is the correct thing here.
Main purpose of this macro on the i915 side is to do short waits on GPU
registers changing post write from spin-locked sections. But there were
rare cases when very short waits were needed from unlocked sections,
shorter than 10us (which is AFAIR what usleep_range documents should be
a lower limit). Which is why non-atomic path was added to the macro.
That path uses preempt_disable/enable so it can use local_clock().
All this may, or may not be, compatible with PREEMPT_RT to start with?
Or question phrased differently, how we should implement the <10us waits
from non-atomic sections under PREEMPT_RT?
> Side note: All of these patches is a collection over time. I personally
> have only a single i7-sandybridge with i915 and here I don't really
> enter all the possible paths here. People report, I patch and look
> around and then they are quiet so I assume that it is working.
>
>>> 0006-drm-i915-Disable-tracing-points-on-PREEMPT_RT.patch
>>
>> If the issue is only with certain trace points why disable all?
>
> It is a class and it is easier that way.
>
>>> 0007-drm-i915-skip-DRM_I915_LOW_LEVEL_TRACEPOINTS-with-NO.patch
>>
>> Didn't quite fully understand, why is this not fixable? Especially thinking
>> if the option of not blanket disabling all tracepoints in the previous
>> patch.
>
> The problem is that you can't acquire that lock from within that
> trace-point on PREEMPT_RT. On !RT it is possible but it is also
> problematic because LOCKDEP does not see possible dead locks unless that
> trace-point is enabled.
Oh I meant could the include ordering problem be fixed differently?
"""
[PATCH 07/10] drm/i915: skip DRM_I915_LOW_LEVEL_TRACEPOINTS with
NOTRACE
The order of the header files is important. If this header file is
included after tracepoint.h was included then the NOTRACE here becomes a
nop. Currently this happens for two .c files which use the tracepoitns
behind DRM_I915_LOW_LEVEL_TRACEPOINTS.
"""
Like these two .c files - can order of includes just be changed in them?
>
> I've been talking to Steven (after
> https://lkml.kernel.org/r/20211214115837.6f33a9b2@gandalf.local.home)
> and he wants to come up with something where you can pass a lock as
> argument to the tracing-API. That way the lock can be acquired before
> the trace event is invoked and lockdep will see it even if the trace
> event is disabled.
> So there is an idea how to get it to work eventually without disabling
> it in the long term.
>
> Making the register a raw_spinlock_t would solve problem immediately but
> I am a little worried given the increased latency in a quick test:
> https://lore.kernel.org/all/20211006164628.s2mtsdd2jdbfyf7g@linutronix.de/
>
> also, this one single hardware but the upper limit atomic-polls is high.
>
>>> 0008-drm-i915-gt-Queue-and-wait-for-the-irq_work-item.patch
>>
>> Not sure about why cond_resched was put between irq_work_queue and
>> irq_work_sync - would it not be like-for-like change to have the two
>> together?
>
> maybe it loops for a while and an additional scheduling would be nice.
>
>> Commit message makes me think _queue already starts the handler on
>> x86 at least.
>
> Yes, irq_work_queue() triggers the IRQ right away on x86,
> irq_work_sync() would wait for it to happen in case it did not happen.
> On architectures which don't provide an IRQ-work interrupt, it is
> delayed to the HZ tick timer interrupt. So this serves also as an
> example in case someone want to copy the code ;)
My question wasn't why is there a need_resched() in there, but why is
the patch:
+ irq_work_queue(&b->irq_work);
cond_resched();
+ irq_work_sync(&b->irq_work);
And not:
+ irq_work_queue(&b->irq_work);
+ irq_work_sync(&b->irq_work);
cond_resched();
To preserve like for like, if my understanding of the commit message was
correct.
>
>>> 0009-drm-i915-gt-Use-spin_lock_irq-instead-of-local_irq_d.patch
>>
>> I think this is okay. The part after the unlock is serialized by the tasklet
>> already.
>>
>> Slight doubt due the comment:
>>
>> local_irq_enable(); /* flush irq_work (e.g. breadcrumb enabling) */
>>
>> Makes me want to think about it harder but not now.
>
> Clark reported it and confirmed that the warning is gone on RT and
> everything appears to work ;)
I will need to think about it harder at some point.
> PREEMPT_RT wise there is no synchronisation vs irq_work other than an
> actual lock (in case it is needed).
Okay, marking as todo/later for me. Need to see if enabling breadcrumbs
earlier than it used to be after this patch makes any difference.
>> Another thing to check is if execlists_context_status_change still needs the
>> atomic notifier chain with this change.
>>
>>> 0010-drm-i915-Drop-the-irqs_disabled-check.patch
>>
>> LGTM.
>
> Do you want me to repost that one?
I think it's up to you whether you go one by one, or repost the whole
series or whatever.
>>> Revert-drm-i915-Depend-on-PREEMPT_RT.patch
>>
>> Okay.
>>
>> And finally for this very patch (the thread I am replying to):
>>
>> Acked-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>
> Thanks.
Np - and I've pushed this one.
Regards,
Tvrtko
More information about the Intel-gfx
mailing list