[PATCH] drm/amdgpu: check flag ring->no_scheduler before usage

Lazar, Lijo lijo.lazar at amd.com
Tue Jan 23 09:00:51 UTC 2024



On 1/21/2024 5:49 AM, vitaly.prosyak at amd.com wrote:
> From: Vitaly Prosyak <vitaly.prosyak at amd.com>
> 
>    The issue started to appear after the following commit
>  11b3b9f461c5c4f700f6c8da202fcc2fd6418e1f (scheduler to variable number
>  of run-queues). The scheduler flag ready (ring->sched.ready) could not be
>  used to validate multiple scenarios, for example, check job is running,
>  gpu_reset, PCI errors etc. The reason is that after GPU reset, the flag
>  is set to true unconditionally even for those rings with an uninitialized scheduler.
>  As a result, we called drm_sched_stop, drm_sched_start for the uninitialized
>  schedule and NULL pointer dereference is occured. For example, the following
>  occurs on Navi10 during GPU reset:
> 
>  [  354.231044] Hardware name: TYAN B8021G88V2HR-2T/S8021GM2NR-2T, BIOS V1.03.B10 04/01/2019
>  [  354.239152] Workqueue: amdgpu-reset-dev drm_sched_job_timedout [gpu_sched]
>  [  354.246047] RIP: 0010:__flush_work.isra.0+0x23a/0x250
>  [  354.251110] Code: 8b 04 25 40 2e 03 00 48 89 44 24 40 48 8b 73 40 8b 4b 30 e9 f9 fe ff ff 40 30 f6 4c 8b 36 e9 37 fe ff ff 0f 0b e9 3a ff ff ff <0f> 0b e9 33 ff ff ff 66
>  66 2e 0f 1f 84 00 00 00 00 00 0f 1f 40 00
>  [  354.269876] RSP: 0018:ffffb234c00e3c20 EFLAGS: 00010246
>  [  354.275121] RAX: 0000000000000011 RBX: ffff9796d9796de0 RCX: 0000000000000000
>  [  354.282271] RDX: 0000000000000001 RSI: 0000000000000000 RDI: ffff9796d9796de0
>  [  354.289420] RBP: ffff9796d9796de0 R08: ffff977780401940 R09: ffffffffa1a5c620
>  [  354.296570] R10: 0000000000000000 R11: 0000000000000000 R12: 0000000000000000
>  [  354.303720] R13: 0000000000000001 R14: ffff9796d97905c8 R15: ffff9796d9790230
>  [  354.310868] FS:  0000000000000000(0000) GS:ffff97865f040000(0000) knlGS:0000000000000000
>  [  354.318963] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>  [  354.324717] CR2: 00007fd5341fca50 CR3: 0000002c27a22000 CR4: 00000000003506f0
>  [  354.324717] CR2: 00007fd5341fca50 CR3: 0000002c27a22000 CR4: 00000000003506f0
>  [  354.331859] Call Trace:
>  [  354.334320]  <TASK>
>  [  354.336433]  ? __flush_work.isra.0+0x23a/0x250
>  [  354.340891]  ? __warn+0x81/0x130
>  [  354.344139]  ? __flush_work.isra.0+0x23a/0x250
>  [  354.348594]  ? report_bug+0x171/0x1a0
>  [  354.352279]  ? handle_bug+0x3c/0x80
>  [  354.355787]  ? exc_invalid_op+0x17/0x70
>  [  354.359635]  ? asm_exc_invalid_op+0x1a/0x20
>  [  354.363844]  ? __flush_work.isra.0+0x23a/0x250
>  [  354.368307]  ? srso_return_thunk+0x5/0x5f
>  [  354.372331]  ? srso_return_thunk+0x5/0x5f
>  [  354.376351]  ? desc_read_finalized_seq+0x1f/0x70
>  [  354.380982]  ? srso_return_thunk+0x5/0x5f
>  [  354.385011]  ? _prb_read_valid+0x20e/0x280
>  [  354.389130]  __cancel_work_timer+0xd3/0x160
>  [  354.393333]  drm_sched_stop+0x46/0x1f0 [gpu_sched]
>  [  354.398143]  amdgpu_device_gpu_recover+0x318/0xca0 [amdgpu]
>  [  354.403995]  ? __drm_err+0x1/0x70 [drm]
>  [  354.407884]  amdgpu_job_timedout+0x151/0x240 [amdgpu]
>  [  354.413279]  drm_sched_job_timedout+0x76/0x100 [gpu_sched]
>  [  354.418787]  process_one_work+0x174/0x340
>  [  354.422816]  worker_thread+0x27b/0x3a0
>  [  354.426586]  ? __pfx_worker_thread+0x10/0x10
>  [  354.430874]  kthread+0xe8/0x120
>  [  354.434030]  ? __pfx_kthread+0x10/0x10
>  [  354.437790]  ret_from_fork+0x34/0x50
>  [  354.441377]  ? __pfx_kthread+0x10/0x10
>  [  354.445139]  ret_from_fork_asm+0x1b/0x30
>  [  354.449079]  </TASK>
>  [  354.451285] ---[ end trace 0000000000000000 ]---
>  [  354.455917] BUG: kernel NULL pointer dereference, address: 0000000000000008
>  [  354.462883] #PF: supervisor read access in kernel mode
>  [  354.468029] #PF: error_code(0x0000) - not-present page
>  [  354.473167] PGD 0 P4D 0
>  [  354.475705] Oops: 0000 [#1] PREEMPT SMP NOPTI
>  [  354.480066] CPU: 1 PID: 11 Comm: kworker/u64:0 Tainted: G        W          6.7.0-991912.1.zuul.e7596ab24dae4bb686e58b0f1e7842da #1
>  [  354.491883] Hardware name: TYAN B8021G88V2HR-2T/S8021GM2NR-2T, BIOS V1.03.B10 04/01/2019
>  [  354.499976] Workqueue: amdgpu-reset-dev drm_sched_job_timedout [gpu_sched]
>  [  354.506855] RIP: 0010:drm_sched_stop+0x61/0x1f0 [gpu_sched]
> 
>   The solution is every place where we check the ready flag and check
>  for ring->no_scheduler. The ready flag serves the purpose in case an initialization
>  is failed, like starting the worker thread, etc.
> 
> Cc: Alex Deucher <alexander.deucher at amd.com>
> Cc: Christian Koenig <christian.koenig at amd.com>
> Signed-off-by: Vitaly Prosyak <vitaly.prosyak at amd.com>
> ---
>  .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c    |  2 ++
>  drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c        |  6 +++---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_device.c         | 14 ++++++++------
>  3 files changed, 13 insertions(+), 9 deletions(-)
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c
> index 899e31e3a5e8..70bbf602df34 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c
> @@ -292,6 +292,8 @@ static int suspend_resume_compute_scheduler(struct amdgpu_device *adev, bool sus
>  
>  		if (!(ring && drm_sched_wqueue_ready(&ring->sched)))
>  			continue;
> +		if (ring->no_scheduler)
> +			continue;
>  

There was a similar patch before -

https://lore.kernel.org/all/7cd37500-8633-4030-aae3-9b532d60b501@amd.com/T/#t

It introduces amdgpu_ring_sched_ready() to cover the above checks.

Thanks,
Lijo

>  		/* stop secheduler and drain ring. */
>  		if (suspend) {
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> index e485dd3357c6..35132aa2c0f4 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> @@ -1678,7 +1678,7 @@ static int amdgpu_debugfs_test_ib_show(struct seq_file *m, void *unused)
>  	for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
>  		struct amdgpu_ring *ring = adev->rings[i];
>  
> -		if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +		if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  			continue;
>  		drm_sched_wqueue_stop(&ring->sched);
>  	}
> @@ -1694,7 +1694,7 @@ static int amdgpu_debugfs_test_ib_show(struct seq_file *m, void *unused)
>  	for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
>  		struct amdgpu_ring *ring = adev->rings[i];
>  
> -		if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +		if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  			continue;
>  		drm_sched_wqueue_start(&ring->sched);
>  	}
> @@ -1916,7 +1916,7 @@ static int amdgpu_debugfs_ib_preempt(void *data, u64 val)
>  
>  	ring = adev->rings[val];
>  
> -	if (!ring || !ring->funcs->preempt_ib ||
> +	if (!ring || !ring->funcs->preempt_ib || ring->no_scheduler ||
>  	    !drm_sched_wqueue_ready(&ring->sched))
>  		return -EINVAL;
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> index 2df14f0e79d8..894b657df1d3 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
> @@ -5052,7 +5052,7 @@ bool amdgpu_device_has_job_running(struct amdgpu_device *adev)
>  	for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  		struct amdgpu_ring *ring = adev->rings[i];
>  
> -		if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +		if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  			continue;
>  
>  		spin_lock(&ring->sched.job_list_lock);
> @@ -5191,8 +5191,10 @@ int amdgpu_device_pre_asic_reset(struct amdgpu_device *adev,
>  	for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  		struct amdgpu_ring *ring = adev->rings[i];
>  
> -		if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +		if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  			continue;
> +		 if (ring->no_scheduler)
> +			 continue;
>  
>  		/* Clear job fence from fence drv to avoid force_completion
>  		 * leave NULL and vm flush fence in fence drv
> @@ -5658,7 +5660,7 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev,
>  		for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  			struct amdgpu_ring *ring = tmp_adev->rings[i];
>  
> -			if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +			if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  				continue;
>  
>  			drm_sched_stop(&ring->sched, job ? &job->base : NULL);
> @@ -5727,7 +5729,7 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev,
>  		for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  			struct amdgpu_ring *ring = tmp_adev->rings[i];
>  
> -			if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +			if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  				continue;
>  
>  			drm_sched_start(&ring->sched, true);
> @@ -6082,7 +6084,7 @@ pci_ers_result_t amdgpu_pci_error_detected(struct pci_dev *pdev, pci_channel_sta
>  		for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  			struct amdgpu_ring *ring = adev->rings[i];
>  
> -			if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +			if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  				continue;
>  
>  			drm_sched_stop(&ring->sched, NULL);
> @@ -6224,7 +6226,7 @@ void amdgpu_pci_resume(struct pci_dev *pdev)
>  	for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>  		struct amdgpu_ring *ring = adev->rings[i];
>  
> -		if (!ring || !drm_sched_wqueue_ready(&ring->sched))
> +		if (!ring || ring->no_scheduler || !drm_sched_wqueue_ready(&ring->sched))
>  			continue;
>  
>  		drm_sched_start(&ring->sched, true);


More information about the amd-gfx mailing list