[PATCH] drm/sched: Re-queue run job worker when drm_sched_entity_pop_job() returns NULL
Rodrigo Vivi
rodrigo.vivi at intel.com
Mon Feb 5 13:33:21 UTC 2024
On Mon, Feb 05, 2024 at 09:44:56AM +0100, Christian König wrote:
> Am 02.02.24 um 22:58 schrieb Rodrigo Vivi:
> > On Tue, Jan 30, 2024 at 08:05:29AM +0100, Christian König wrote:
> > > Am 30.01.24 um 04:04 schrieb Matthew Brost:
> > > > Rather then loop over entities until one with a ready job is found,
> > > > re-queue the run job worker when drm_sched_entity_pop_job() returns NULL.
> > > >
> > > > Fixes: 6dbd9004a55 ("drm/sched: Drain all entities in DRM sched run job worker")
> > First of all there's a small typo in this Fixes tag that needs to be fixed.
> > The correct one is:
> >
> > Fixes: 66dbd9004a55 ("drm/sched: Drain all entities in DRM sched run job worker")
Cc: Dave Airlie <airlied at redhat.com>
> >
> > But I couldn't apply this right now in any of our drm-tip trees because it
> > is not clear where this is coming from originally.
> >
> > likely amd tree?!
>
> No, this comes from Matthews work on the DRM scheduler.
>
> Matthews patches were most likely merged through drm-misc.
the original is not there in drm-misc-next.
it looks like Dave had taken that one directly to drm-next.
So we either need the drm-misc maintainers to have a backmerge or
Dave to take this through the drm-fixes directly.
>
> Regards,
> Christian.
>
> >
> > > > Signed-off-by: Matthew Brost <matthew.brost at intel.com>
> > > Reviewed-by: Christian König <christian.koenig at amd.com>
> > Christian, if this came from the amd, could you please apply it there and
> > propagate through your fixes flow?
> >
> > Thanks,
> > Rodrigo.
> >
> > > > ---
> > > > drivers/gpu/drm/scheduler/sched_main.c | 15 +++++++++------
> > > > 1 file changed, 9 insertions(+), 6 deletions(-)
> > > >
> > > > diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
> > > > index 8acbef7ae53d..7e90c9f95611 100644
> > > > --- a/drivers/gpu/drm/scheduler/sched_main.c
> > > > +++ b/drivers/gpu/drm/scheduler/sched_main.c
> > > > @@ -1178,21 +1178,24 @@ static void drm_sched_run_job_work(struct work_struct *w)
> > > > struct drm_sched_entity *entity;
> > > > struct dma_fence *fence;
> > > > struct drm_sched_fence *s_fence;
> > > > - struct drm_sched_job *sched_job = NULL;
> > > > + struct drm_sched_job *sched_job;
> > > > int r;
> > > > if (READ_ONCE(sched->pause_submit))
> > > > return;
> > > > /* Find entity with a ready job */
> > > > - while (!sched_job && (entity = drm_sched_select_entity(sched))) {
> > > > - sched_job = drm_sched_entity_pop_job(entity);
> > > > - if (!sched_job)
> > > > - complete_all(&entity->entity_idle);
> > > > - }
> > > > + entity = drm_sched_select_entity(sched);
> > > > if (!entity)
> > > > return; /* No more work */
> > > > + sched_job = drm_sched_entity_pop_job(entity);
> > > > + if (!sched_job) {
> > > > + complete_all(&entity->entity_idle);
> > > > + drm_sched_run_job_queue(sched);
> > > > + return;
> > > > + }
> > > > +
> > > > s_fence = sched_job->s_fence;
> > > > atomic_add(sched_job->credits, &sched->credit_count);
>
More information about the dri-devel
mailing list