[PATCH] drm/sched: add run job trace

Christian König christian.koenig at amd.com
Fri Mar 13 10:17:52 UTC 2020


Am 13.03.20 um 10:56 schrieb Lucas Stach:
> From: Robert Beckett <bob.beckett at collabora.com>
>
> Add a new trace event to show when jobs are run on the HW.
>
> Signed-off-by: Robert Beckett <bob.beckett at collabora.com>
> Signed-off-by: Lucas Stach <l.stach at pengutronix.de>

There is also the scheduled fence we could used for this, but this trace 
point adds a few extra fields which might be useful.

Acked-by: Christian König <christian.koenig at amd.com>

> ---
>   .../gpu/drm/scheduler/gpu_scheduler_trace.h   | 27 +++++++++++++++++++
>   drivers/gpu/drm/scheduler/sched_main.c        |  1 +
>   2 files changed, 28 insertions(+)
>
> diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> index d79086498aff..877ce9b127f1 100644
> --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> @@ -59,6 +59,33 @@ TRACE_EVENT(drm_sched_job,
>   		      __entry->job_count, __entry->hw_job_count)
>   );
>   
> +TRACE_EVENT(drm_run_job,
> +	    TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
> +	    TP_ARGS(sched_job, entity),
> +	    TP_STRUCT__entry(
> +			     __field(struct drm_sched_entity *, entity)
> +			     __field(struct dma_fence *, fence)
> +			     __field(const char *, name)
> +			     __field(uint64_t, id)
> +			     __field(u32, job_count)
> +			     __field(int, hw_job_count)
> +			     ),
> +
> +	    TP_fast_assign(
> +			   __entry->entity = entity;
> +			   __entry->id = sched_job->id;
> +			   __entry->fence = &sched_job->s_fence->finished;
> +			   __entry->name = sched_job->sched->name;
> +			   __entry->job_count = spsc_queue_count(&entity->job_queue);
> +			   __entry->hw_job_count = atomic_read(
> +				   &sched_job->sched->hw_rq_count);
> +			   ),
> +	    TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d",
> +		      __entry->entity, __entry->id,
> +		      __entry->fence, __entry->name,
> +		      __entry->job_count, __entry->hw_job_count)
> +);
> +
>   TRACE_EVENT(drm_sched_process_job,
>   	    TP_PROTO(struct drm_sched_fence *fence),
>   	    TP_ARGS(fence),
> diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
> index 71ce6215956f..34231b7163cc 100644
> --- a/drivers/gpu/drm/scheduler/sched_main.c
> +++ b/drivers/gpu/drm/scheduler/sched_main.c
> @@ -773,6 +773,7 @@ static int drm_sched_main(void *param)
>   		atomic_inc(&sched->hw_rq_count);
>   		drm_sched_job_begin(sched_job);
>   
> +		trace_drm_run_job(sched_job, entity);
>   		fence = sched->ops->run_job(sched_job);
>   		drm_sched_fence_scheduled(s_fence);
>   



More information about the dri-devel mailing list