[PATCH] drm/amdgpu: fix the null pointer to get timeline by scheduler fence
Huang Rui
ray.huang at amd.com
Wed Aug 8 11:28:14 UTC 2018
On Wed, Aug 08, 2018 at 03:10:07PM +0800, Koenig, Christian wrote:
> Yeah that is a known issue, but this solution is not correct either.
>
> See the scheduler where the job is execute on is simply not determined
> yet when we want to trace it.
>
> So using the scheduler name from the entity is wrong as well.
>
> We should probably move the reschedule from drm_sched_entity_push_job()
> to drm_sched_job_init() to fix that.
Could you please explain why move reschedule along can fix the issue.
Seemingly, only s_fence's sched is written to entity rq's sched, it can
avoid the issue.
sched_job->s_fence->sched = entity->rq->sched
Thanks,
Ray
>
> I will prepare a patch for that today,
> Christian.
>
> Am 08.08.2018 um 09:05 schrieb Huang Rui:
> > We won't initialize fence scheduler in drm_sched_fence_create() anymore, so it
> > will refer null fence scheduler if open trace event to get the timeline name.
> > Actually, it is the scheduler name from the entity, so add a macro to replace
> > legacy getting timeline name by job.
> >
> > [ 212.844281] BUG: unable to handle kernel NULL pointer dereference at 0000000000000018
> > [ 212.852401] PGD 8000000427c13067 P4D 8000000427c13067 PUD 4235fc067 PMD 0
> > [ 212.859419] Oops: 0000 [#1] SMP PTI
> > [ 212.862981] CPU: 4 PID: 1520 Comm: amdgpu_test Tainted: G OE 4.18.0-rc1-custom #1
> > [ 212.872194] Hardware name: Gigabyte Technology Co., Ltd. Z170XP-SLI/Z170XP-SLI-CF, BIOS F20 11/04/2016
> > [ 212.881704] RIP: 0010:drm_sched_fence_get_timeline_name+0x2b/0x30 [gpu_sched]
> > [ 212.888948] Code: 1f 44 00 00 48 8b 47 08 48 3d c0 b1 4f c0 74 13 48 83 ef 60 48 3d 60 b1 4f c0 b8 00 00 00 00 48 0f 45 f8 48 8b 87 e0 00 00 00 <48> 8b 40 18 c3 0f 1f 44 00 00 b8 01 00 00 00 c3 0f 1f 44 00 00 0f
> > [ 212.908162] RSP: 0018:ffffa3ed81f27af0 EFLAGS: 00010246
> > [ 212.913483] RAX: 0000000000000000 RBX: 0000000000070034 RCX: ffffa3ed81f27da8
> > [ 212.920735] RDX: ffff8f24ebfb5460 RSI: ffff8f24e40d3c00 RDI: ffff8f24ebfb5400
> > [ 212.928008] RBP: ffff8f24e40d3c00 R08: 0000000000000000 R09: ffffffffae4deafc
> > [ 212.935263] R10: ffffffffada000ed R11: 0000000000000001 R12: ffff8f24e891f898
> > [ 212.942558] R13: 0000000000000000 R14: ffff8f24ebc46000 R15: ffff8f24e3de97a8
> > [ 212.949796] FS: 00007ffff7fd2700(0000) GS:ffff8f24fed00000(0000) knlGS:0000000000000000
> > [ 212.958047] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> > [ 212.963921] CR2: 0000000000000018 CR3: 0000000423422003 CR4: 00000000003606e0
> > [ 212.971201] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
> > [ 212.978482] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
> > [ 212.985720] Call Trace:
> > [ 212.988236] trace_event_raw_event_amdgpu_cs_ioctl+0x4c/0x170 [amdgpu]
> > [ 212.994904] ? amdgpu_ctx_add_fence+0xa9/0x110 [amdgpu]
> > [ 213.000246] ? amdgpu_job_free_resources+0x4b/0x70 [amdgpu]
> > [ 213.005944] amdgpu_cs_ioctl+0x16d1/0x1b50 [amdgpu]
> > [ 213.010920] ? amdgpu_cs_find_mapping+0xf0/0xf0 [amdgpu]
> > [ 213.016354] drm_ioctl_kernel+0x8a/0xd0 [drm]
> > [ 213.020794] ? recalc_sigpending+0x17/0x50
> > [ 213.024965] drm_ioctl+0x2d7/0x390 [drm]
> > [ 213.028979] ? amdgpu_cs_find_mapping+0xf0/0xf0 [amdgpu]
> > [ 213.034366] ? do_signal+0x36/0x700
> > [ 213.037928] ? signal_wake_up_state+0x15/0x30
> > [ 213.042375] amdgpu_drm_ioctl+0x46/0x80 [amdgpu]
> >
> > Signed-off-by: Huang Rui <ray.huang at amd.com>
> > ---
> > drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c | 2 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h | 10 ++++++----
> > 2 files changed, 7 insertions(+), 5 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> > index e12871d..be01e1b 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> > @@ -1247,7 +1247,7 @@ static int amdgpu_cs_submit(struct amdgpu_cs_parser *p,
> >
> > amdgpu_job_free_resources(job);
> >
> > - trace_amdgpu_cs_ioctl(job);
> > + trace_amdgpu_cs_ioctl(job, entity);
> > amdgpu_vm_bo_trace_cs(&fpriv->vm, &p->ticket);
> > priority = job->base.s_priority;
> > drm_sched_entity_push_job(&job->base, entity);
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> > index 8c2dab2..25cdcb7 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> > @@ -36,6 +36,8 @@
> >
> > #define AMDGPU_JOB_GET_TIMELINE_NAME(job) \
> > job->base.s_fence->finished.ops->get_timeline_name(&job->base.s_fence->finished)
> > +#define AMDGPU_GET_SCHED_NAME(entity) \
> > + (entity->rq->sched->name)
> >
> > TRACE_EVENT(amdgpu_mm_rreg,
> > TP_PROTO(unsigned did, uint32_t reg, uint32_t value),
> > @@ -161,11 +163,11 @@ TRACE_EVENT(amdgpu_cs,
> > );
> >
> > TRACE_EVENT(amdgpu_cs_ioctl,
> > - TP_PROTO(struct amdgpu_job *job),
> > - TP_ARGS(job),
> > + TP_PROTO(struct amdgpu_job *job, struct drm_sched_entity *entity),
> > + TP_ARGS(job, entity),
> > TP_STRUCT__entry(
> > __field(uint64_t, sched_job_id)
> > - __string(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
> > + __string(timeline, AMDGPU_GET_SCHED_NAME(entity))
> > __field(unsigned int, context)
> > __field(unsigned int, seqno)
> > __field(struct dma_fence *, fence)
> > @@ -175,7 +177,7 @@ TRACE_EVENT(amdgpu_cs_ioctl,
> >
> > TP_fast_assign(
> > __entry->sched_job_id = job->base.id;
> > - __assign_str(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
> > + __assign_str(timeline, AMDGPU_GET_SCHED_NAME(entity))
> > __entry->context = job->base.s_fence->finished.context;
> > __entry->seqno = job->base.s_fence->finished.seqno;
> > __entry->ring_name = to_amdgpu_ring(job->base.sched)->name;
>
More information about the amd-gfx
mailing list