[PATCH] drm/amdgpu: trace amdgpu_job fence details

Andres Rodriguez andresx7 at gmail.com
Thu Feb 23 20:48:24 UTC 2017



On 2017-02-23 02:46 PM, Andres Rodriguez wrote:
>
>
> On 2017-02-23 03:20 AM, Christian König wrote:
>> Am 23.02.2017 um 00:47 schrieb Andres Rodriguez:
>>> This trace is intended to provide the required information to associate
>>> the completion of an amdgpu_job with its corresponding dma_fence_*
>>> tracepoints.
>>>
>>> Signed-off-by: Andres Rodriguez <andresx7 at gmail.com>
>>> ---
>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_job.c   |  2 ++
>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h | 22 ++++++++++++++++++++++
>>>   2 files changed, 24 insertions(+)
>>>
>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 
>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>> index 86a1242..84a04e4 100644
>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
>>> @@ -177,6 +177,8 @@ static struct dma_fence *amdgpu_job_run(struct 
>>> amd_sched_job *sched_job)
>>>       /* if gpu reset, hw fence will be replaced here */
>>>       dma_fence_put(job->fence);
>>>       job->fence = dma_fence_get(fence);
>>> +    trace_amdgpu_job_attach_fence(job);
>>> +
>>>       amdgpu_job_free_resources(job);
>>>       return fence;
>>>   }
>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 
>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
>>> index a18ae1e..0a61ed9 100644
>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
>>> @@ -147,6 +147,28 @@ TRACE_EVENT(amdgpu_sched_run_job,
>>>                 __entry->fence, __entry->ring_name, __entry->num_ibs)
>>>   );
>>>   +TRACE_EVENT(amdgpu_job_attach_fence,
>>> +        TP_PROTO(struct amdgpu_job *job),
>>> +        TP_ARGS(job),
>>> +        TP_STRUCT__entry(
>>> +                 __field(struct amdgpu_device *, adev)
>>> +                 __field(struct amd_sched_job *, sched_job)
>>
>> Neither add adev nor sched_job pointer to your trace point.
> Would it be okay then to grab a reference to the scheduler fence in 
> amdgpu_job_run and print the sched fence pointer here?
>
> That would give us enough info to correlate other trace points like 
> amdgpu_cs_ioctl aor amdgpu_sched_run_job to the respective dma_fence 
> tracepoints.
>
>
>>
>> The adev pointer is completely meaningless in the logs and the 
>> scheduler job might already be freed when the printk is called.
Hey Christian,

I was just double checking the lifetime of the scheduler job, and it 
should still be valid during amdgpu_job_run. The lifetime of the job is 
tied to the scheduler finished fence which won't be signaled until 
amdgpu_job_run finishes even if the HW fence has signaled.

This is assuming that the trace printks are synchronous.

So I think printing the job and dropping adev should be okay if the 
above holds true, unless I'm missing something else.

>>
>> Just checked the source and a couple of other trace points get this 
>> horrible wrong as well, so that isn't your fault. Going to put it on 
>> my todo to fix those.
>>
>> Regards,
>> Christian.
>>
>>> +                 __string(timeline, 
>>> job->fence->ops->get_timeline_name(job->fence))
>>> +                 __field(unsigned int, context)
>>> +                 __field(unsigned int, seqno)
>>> +                 ),
>>> +
>>> +        TP_fast_assign(
>>> +               __entry->adev = job->adev;
>>> +               __entry->sched_job = &job->base;
>>> +               __assign_str(timeline, 
>>> job->fence->ops->get_timeline_name(job->fence))
>>> +               __entry->context = job->fence->context;
>>> +               __entry->seqno = job->fence->seqno;
>>> +               ),
>>> +        TP_printk("adev=%p, sched_job=%p, timeline:%s, context:%u, 
>>> seqno:%u",
>>> +              __entry->adev, __entry->sched_job,
>>> +              __get_str(timeline), __entry->context, __entry->seqno)
>>> +);
>>>     TRACE_EVENT(amdgpu_vm_grab_id,
>>>           TP_PROTO(struct amdgpu_vm *vm, int ring, struct amdgpu_job 
>>> *job),
>>
>>
>



More information about the amd-gfx mailing list