[PATCH 2/2] drm/xe: Don't free job in TDR
Matthew Auld
matthew.auld at intel.com
Thu Oct 3 14:15:02 UTC 2024
On 03/10/2024 15:05, Matthew Brost wrote:
> On Thu, Oct 03, 2024 at 08:06:24AM +0100, Matthew Auld wrote:
>> On 03/10/2024 01:16, Matthew Brost wrote:
>>> Freeing job in TDR is not safe as TDR can pass the run_job thread
>>> resulting in UAF. It is only safe for free job to naturally be called by
>>> the scheduler. Rather free job in TDR, add to pending list.
>>
>> s/Rather free/Rather than free/
>> ?
>>
>
> Yes, will fix.
>
>>>
>>> Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/2811
>>> Cc: Matthew Auld <matthew.auld at intel.com>
>>> Fixes: e275d61c5f3f ("drm/xe/guc: Handle timing out of signaled jobs gracefully")
>>> Signed-off-by: Matthew Brost <matthew.brost at intel.com>
>>
>> I think we still have the other issue with fence signalling in run_job.
>>
>
> I think this actually ok given free_job as owns a ref to job->fence and
> free_job now must run after run_job - that is why I didn't include this
> change in this patch. But I also agree a better design would be move the
> dma_fence_get from run_job to arm - I will do that in a follow up.
Here I mean the race in run_job() itself, before we hand over the fence
to the scheduler. i.e do the dma_fence_get() before the submission part
like in:
https://patchwork.freedesktop.org/patch/615249/?series=138921&rev=1.
>
> Matt
>
>> Reviewed-by: Matthew Auld <matthew.auld at intel.com>
>>
>>> ---
>>> drivers/gpu/drm/xe/xe_guc_submit.c | 7 +++++--
>>> 1 file changed, 5 insertions(+), 2 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c
>>> index 80062e1d3f66..9ecd1661c1b5 100644
>>> --- a/drivers/gpu/drm/xe/xe_guc_submit.c
>>> +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
>>> @@ -1106,10 +1106,13 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
>>> /*
>>> * TDR has fired before free job worker. Common if exec queue
>>> - * immediately closed after last fence signaled.
>>> + * immediately closed after last fence signaled. Add back to pending
>>> + * list so job can be freed and kick scheduler ensuring free job is not
>>> + * lost.
>>> */
>>> if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &job->fence->flags)) {
>>> - guc_exec_queue_free_job(drm_job);
>>> + xe_sched_add_pending_job(sched, job);
>>> + xe_sched_submission_start(sched);
>>> return DRM_GPU_SCHED_STAT_NOMINAL;
>>> }
More information about the Intel-xe
mailing list