[PATCH] drm/amdkfd: Don't call mmput from MMU notifier callback
Philip Yang
yangp at amd.com
Tue Jun 24 15:00:56 UTC 2025
On 2025-06-23 18:18, Chen, Xiaogang wrote:
>
>
> On 6/23/2025 11:59 AM, Philip Yang wrote:
>> If the process is exiting, the mmput inside mmu notifier callback from
>> compactd or fork or numa balancing could release the last reference
>> of mm struct to call exit_mmap and free_pgtable, this triggers deadlock
>> with below backtrace.
>>
>> The deadlock will leak kfd process as mmu notifier release is not called
>> and cause VRAM leaking.
>>
>> The fix is to take mm reference mmget_non_zero when adding prange to the
>> deferred list to pair with mmput in deferred list work.
>>
>> The backtrace of hung task:
>>
>> INFO: task python:348105 blocked for more than 64512 seconds.
>> Call Trace:
>> __schedule+0x1c3/0x550
>> schedule+0x46/0xb0
>> rwsem_down_write_slowpath+0x24b/0x4c0
>> unlink_anon_vmas+0xb1/0x1c0
>> free_pgtables+0xa9/0x130
>> exit_mmap+0xbc/0x1a0
>> mmput+0x5a/0x140
>> svm_range_cpu_invalidate_pagetables+0x2b/0x40 [amdgpu]
>> mn_itree_invalidate+0x72/0xc0
>> __mmu_notifier_invalidate_range_start+0x48/0x60
>> try_to_unmap_one+0x10fa/0x1400
>> rmap_walk_anon+0x196/0x460
>> try_to_unmap+0xbb/0x210
>> migrate_page_unmap+0x54d/0x7e0
>> migrate_pages_batch+0x1c3/0xae0
>> migrate_pages_sync+0x98/0x240
>> migrate_pages+0x25c/0x520
>> compact_zone+0x29d/0x590
>> compact_zone_order+0xb6/0xf0
>> try_to_compact_pages+0xbe/0x220
>> __alloc_pages_direct_compact+0x96/0x1a0
>> __alloc_pages_slowpath+0x410/0x930
>> __alloc_pages_nodemask+0x3a9/0x3e0
>> do_huge_pmd_anonymous_page+0xd7/0x3e0
>> __handle_mm_fault+0x5e3/0x5f0
>> handle_mm_fault+0xf7/0x2e0
>> hmm_vma_fault.isra.0+0x4d/0xa0
>> walk_pmd_range.isra.0+0xa8/0x310
>> walk_pud_range+0x167/0x240
>> walk_pgd_range+0x55/0x100
>> __walk_page_range+0x87/0x90
>> walk_page_range+0xf6/0x160
>> hmm_range_fault+0x4f/0x90
>> amdgpu_hmm_range_get_pages+0x123/0x230 [amdgpu]
>> amdgpu_ttm_tt_get_user_pages+0xb1/0x150 [amdgpu]
>> init_user_pages+0xb1/0x2a0 [amdgpu]
>> amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu+0x543/0x7d0 [amdgpu]
>> kfd_ioctl_alloc_memory_of_gpu+0x24c/0x4e0 [amdgpu]
>> kfd_ioctl+0x29d/0x500 [amdgpu]
>>
>> Fixes: fa582c6f3684 ("drm/amdkfd: Use mmget_not_zero in MMU notifier")
>> Signed-off-by: Philip Yang<Philip.Yang at amd.com>
>> ---
>> drivers/gpu/drm/amd/amdkfd/kfd_svm.c | 23 +++++++++++------------
>> 1 file changed, 11 insertions(+), 12 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> index 10d1276f8e1c..5fe92f9a1ce1 100644
>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> @@ -2392,15 +2392,17 @@ svm_range_add_list_work(struct svm_range_list *svms, struct svm_range *prange,
>> prange->work_item.op != SVM_OP_UNMAP_RANGE)
>> prange->work_item.op = op;
>> } else {
>> - prange->work_item.op = op;
>> -
>> - /* Pairs with mmput in deferred_list_work */
>> - mmget(mm);
>> - prange->work_item.mm = mm;
>> - list_add_tail(&prange->deferred_list,
>> - &prange->svms->deferred_range_list);
>> - pr_debug("add prange 0x%p [0x%lx 0x%lx] to work list op %d\n",
>> - prange, prange->start, prange->last, op);
>> + /* Pairs with mmput in deferred_list_work.
>> + * If process is exiting and mm is gone, don't update mmu notifier.
>> + */
>> + if (mmget_not_zero(mm)) {
> If process is exiting we not need do schedule_deferred_list_work
> neither. I think this part code need be reorganized with
> mmget_not_zero(mm) check.
yes, that is right, we could change svm_range_add_list_work to return
true only if new work_item is added to the deferred list, and schedule
deferred work. This optimization could be another patch, not related to
the deadlock issue.
>> + prange->work_item.mm = mm;
>> + prange->work_item.op = op;
>> + list_add_tail(&prange->deferred_list,
>> + &prange->svms->deferred_range_list);
>> + pr_debug("add prange 0x%p [0x%lx 0x%lx] to work list op %d\n",
>> + prange, prange->start, prange->last, op);
>> + }
>> }
>> spin_unlock(&svms->deferred_list_lock);
>> }
>> @@ -2568,8 +2570,6 @@ svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni,
>>
>> if (range->event == MMU_NOTIFY_RELEASE)
>> return true;
>> - if (!mmget_not_zero(mni->mm))
>> - return true;
>
> Why remove mmget_not_zero(mni->mm) /mmput(mni->mm) here? I think they
> are for different purpose from mmget_not_zero(mm) at
> svm_range_add_list_work.
>
Yes, the mmget_non_zero/mmput inside MMU notifier was added to fix
another issue that deferred work access invalid mm, to ensure deferred
work_item.mm hold mm reference. But in the backtrace case, mmput inside
MMU notifier callback release the last reference and cause deadlock.
Regards,
Philip
> Regards
>
> Xiaogang
>
>>
>> start = mni->interval_tree.start;
>> last = mni->interval_tree.last;
>> @@ -2596,7 +2596,6 @@ svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni,
>> }
>>
>> svm_range_unlock(prange);
>> - mmput(mni->mm);
>>
>> return true;
>> }
More information about the amd-gfx
mailing list