[PATCH v3] drm/panthor: Make the timeout per-queue instead of per-job

Boris Brezillon boris.brezillon at collabora.com
Mon Apr 14 16:56:23 UTC 2025


On Mon, 14 Apr 2025 17:44:27 +0100
Ashley Smith <ashley.smith at collabora.com> wrote:

> On Fri, 11 Apr 2025 16:51:52 +0100 Steven Price  wrote:
>  > Hi Ashley, 
>  >  
>  > On 10/04/2025 13:57, Ashley Smith wrote:   
>  > > The timeout logic provided by drm_sched leads to races when we try 
>  > > to suspend it while the drm_sched workqueue queues more jobs. Let's 
>  > > overhaul the timeout handling in panthor to have our own delayed work 
>  > > that's resumed/suspended when a group is resumed/suspended. When an 
>  > > actual timeout occurs, we call drm_sched_fault() to report it 
>  > > through drm_sched, still. But otherwise, the drm_sched timeout is 
>  > > disabled (set to MAX_SCHEDULE_TIMEOUT), which leaves us in control of 
>  > > how we protect modifications on the timer. 
>  > > 
>  > > One issue seems to be when we call drm_sched_suspend_timeout() from 
>  > > both queue_run_job() and tick_work() which could lead to races due to 
>  > > drm_sched_suspend_timeout() not having a lock. Another issue seems to 
>  > > be in queue_run_job() if the group is not scheduled, we suspend the 
>  > > timeout again which undoes what drm_sched_job_begin() did when calling 
>  > > drm_sched_start_timeout(). So the timeout does not reset when a job 
>  > > is finished. 
>  > > 
>  > > Co-developed-by: Boris Brezillon boris.brezillon at collabora.com> 
>  > > Signed-off-by: Boris Brezillon boris.brezillon at collabora.com> 
>  > > Tested-by: Daniel Stone daniels at collabora.com> 
>  > > Fixes: de8548813824 ("drm/panthor: Add the scheduler logical block") 
>  > > Signed-off-by: Ashley Smith ashley.smith at collabora.com> 
>  > > --- 
>  > >  drivers/gpu/drm/panthor/panthor_sched.c | 244 +++++++++++++++++------- 
>  > >  1 file changed, 177 insertions(+), 67 deletions(-) 
>  > > 
>  > > diff --git a/drivers/gpu/drm/panthor/panthor_sched.c b/drivers/gpu/drm/panthor/panthor_sched.c 
>  > > index 446ec780eb4a..32f5a75bc4f6 100644 
>  > > --- a/drivers/gpu/drm/panthor/panthor_sched.c 
>  > > +++ b/drivers/gpu/drm/panthor/panthor_sched.c   
>  >  
>  > [...] 
>  >    
>  > > @@ -2727,8 +2784,17 @@ void panthor_sched_suspend(struct panthor_device *ptdev) 
>  > >               * automatically terminate all active groups, so let's 
>  > >               * force the state to halted here. 
>  > >               */ 
>  > > -            if (csg_slot->group->state != PANTHOR_CS_GROUP_TERMINATED) 
>  > > +            if (csg_slot->group->state != PANTHOR_CS_GROUP_TERMINATED) { 
>  > >                  csg_slot->group->state = PANTHOR_CS_GROUP_TERMINATED; 
>  > > + 
>  > > +                /* Reset the queue slots manually if the termination 
>  > > +                * request failed. 
>  > > +                */ 
>  > > +                for (i = 0; i queue_count; i++) { 
>  > > +                    if (group->queues[i]) 
>  > > +                        cs_slot_reset_locked(ptdev, csg_id, i); 
>  > > +                } 
>  > > +            } 
>  > >              slot_mask &= ~BIT(csg_id); 
>  > >          } 
>  > >      }   
>  >  
>  > So this seems to be the only change from v2 (a changelog can be 
>  > helpful!). And I'm not convinced it belongs in this patch? It's not just 
>  > "[making] the timeout per-queue instead of per-job". 
>  >  
>  > I haven't dug through the details, but I think this belongs in a 
>  > separate patch.

Actually, it's related, but I agree it's not clear: we call
cs_slot_reset_locked(), but the thing we're really interested in is the
cancellation of the timeout work. Before the timeout changes, the
timeout work was part of drm_sched and was cancelled inside
drm_sched_stop(). Now, maybe we do need to reset the CS slot
regardless, in which case it might make sense to have that done in a
separate fix happening before the timeout changes.


More information about the dri-devel mailing list