[PATCH] drm/xe/guc_submit: fix UAF in run_job()
Matthew Brost
matthew.brost at intel.com
Fri Sep 20 19:07:36 UTC 2024
On Fri, Sep 20, 2024 at 01:38:07PM +0100, Matthew Auld wrote:
> The initial kref from dma_fence_init() should match up with whatever
> signals the fence, however here we are submitting the job first to the
> hw and only then grabbing the extra ref and even then we touch some
> fence state before this. This might be too late if the fence is
> signalled before we can grab the extra ref. Rather always grab the
> refcount early before we do the submission part.
>
I think I see the race. Let me make sure I understand.
Current flow:
1. guc_exec_queue_run_job enters
2. guc_exec_queue_run_job submits job to hardware
3. job finishes on hardware
4. irq handler for job completion fires, signals job->fence, does last
put on job->fence freeing the memory
5. guc_exec_queue_run_job takes a ref job->fence and BOOM UAF
The extra ref between steps 1/2 dropped after 5 prevents this. Is that
right?
Assuming my understanding is correct:
Reviewed-by: Matthew Brost <matthew.brost at intel.com>
> Closes: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/2811
> Signed-off-by: Matthew Auld <matthew.auld at intel.com>
> Fixes: dd08ebf6c352 ("drm/xe: Introduce a new DRM driver for Intel GPUs")
> Cc: Matthew Brost <matthew.brost at intel.com>
> Cc: <stable at vger.kernel.org> # v6.8+
> ---
> drivers/gpu/drm/xe/xe_guc_submit.c | 13 ++++++++++---
> 1 file changed, 10 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c
> index fbbe6a487bbb..b33f3d23a068 100644
> --- a/drivers/gpu/drm/xe/xe_guc_submit.c
> +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
> @@ -766,12 +766,15 @@ guc_exec_queue_run_job(struct drm_sched_job *drm_job)
> struct xe_guc *guc = exec_queue_to_guc(q);
> struct xe_device *xe = guc_to_xe(guc);
> bool lr = xe_exec_queue_is_lr(q);
> + struct dma_fence *fence;
>
> xe_assert(xe, !(exec_queue_destroyed(q) || exec_queue_pending_disable(q)) ||
> exec_queue_banned(q) || exec_queue_suspended(q));
>
> trace_xe_sched_job_run(job);
>
> + dma_fence_get(job->fence);
> +
> if (!exec_queue_killed_or_banned_or_wedged(q) && !xe_sched_job_is_error(job)) {
> if (!exec_queue_registered(q))
> register_exec_queue(q);
> @@ -782,12 +785,16 @@ guc_exec_queue_run_job(struct drm_sched_job *drm_job)
>
> if (lr) {
> xe_sched_job_set_error(job, -EOPNOTSUPP);
> - return NULL;
> + fence = NULL;
> } else if (test_and_set_bit(JOB_FLAG_SUBMIT, &job->fence->flags)) {
> - return job->fence;
> + fence = job->fence;
> } else {
> - return dma_fence_get(job->fence);
> + fence = dma_fence_get(job->fence);
> }
> +
> + dma_fence_put(job->fence);
> +
> + return fence;
> }
>
> static void guc_exec_queue_free_job(struct drm_sched_job *drm_job)
> --
> 2.46.0
>
More information about the Intel-xe
mailing list