[PATCH 12/18] drm/xe/eudebug: implement userptr_vma access

Hajda, Andrzej andrzej.hajda at intel.com
Mon Oct 21 09:54:30 UTC 2024


W dniu 20.10.2024 o 20:16, Matthew Brost pisze:
> On Tue, Oct 01, 2024 at 05:43:00PM +0300, Mika Kuoppala wrote:
>> From: Andrzej Hajda <andrzej.hajda at intel.com>
>>
>> Debugger needs to read/write program's vmas including userptr_vma.
>> Since hmm_range_fault is used to pin userptr vmas, it is possible
>> to map those vmas from debugger context.
>>
>> v2: pin pages vs notifier, move to vm.c (Matthew)
>>
>> Signed-off-by: Andrzej Hajda <andrzej.hajda at intel.com>
>> Signed-off-by: Maciej Patelczyk <maciej.patelczyk at intel.com>
>> Signed-off-by: Mika Kuoppala <mika.kuoppala at linux.intel.com>
>> Reviewed-by: Jonathan Cavitt <jonathan.cavitt at intel.com>
>> ---
>>   drivers/gpu/drm/xe/xe_eudebug.c |  2 +-
>>   drivers/gpu/drm/xe/xe_vm.c      | 47 +++++++++++++++++++++++++++++++++
>>   drivers/gpu/drm/xe/xe_vm.h      |  3 +++
>>   3 files changed, 51 insertions(+), 1 deletion(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_eudebug.c b/drivers/gpu/drm/xe/xe_eudebug.c
>> index edad6d533d0b..b09d7414cfe3 100644
>> --- a/drivers/gpu/drm/xe/xe_eudebug.c
>> +++ b/drivers/gpu/drm/xe/xe_eudebug.c
>> @@ -3023,7 +3023,7 @@ static int xe_eudebug_vma_access(struct xe_vma *vma, u64 offset,
>>   		return ret;
>>   	}
>>   
>> -	return -EINVAL;
>> +	return xe_uvma_access(to_userptr_vma(vma), offset, buf, bytes, write);
>>   }
>>   
>>   static int xe_eudebug_vm_access(struct xe_vm *vm, u64 offset,
>> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
>> index a836dfc5a86f..5f891e76993b 100644
>> --- a/drivers/gpu/drm/xe/xe_vm.c
>> +++ b/drivers/gpu/drm/xe/xe_vm.c
>> @@ -3421,3 +3421,50 @@ void xe_vm_snapshot_free(struct xe_vm_snapshot *snap)
>>   	}
>>   	kvfree(snap);
>>   }
>> +
>> +int xe_uvma_access(struct xe_userptr_vma *uvma, u64 offset,
>> +		   void *buf, u64 len, bool write)
>> +{
> 
> Maybe dump question but are we overthinking this here?
> 
> Can we just use kthread_use_mm, copy_to_user, copy_from_user?
> 
> If not then my previous comments still apply here.

This function is called from debugger process context and kthread_use_mm 
is allowed only from kthread. Spawning kthread just for this is an 
option but looks odd and suboptimal, could be kind of last resort, or not?

Another options:
1. Keep reference to remote task in xe_userptr and use 
access_process_vm(up->task, ...).

2. Pass xe_eudebug.target_task reference down from eudebug framework to 
this helper and use access_process_vm. Current call chain is:
__xe_eudebug_vm_access - has access to xe_eudebug.target_task
->__vm_read_write
--->xe_eudebug_vm_access
---->xe_eudebug_vm_access
----->xe_eudebug_vma_access
------>xe_vm_userptr_access
So to achieve this multiple changes are required, but maybe it is valid 
path to go?
One potential issue with 1 and 2 is that multiple UMD tests were failing 
when access_process_vm/access_remote_vm were used, they were not 
investigated as this approach was dropped due to different reasons.

3. Continue approach from this patch, but with corrected page iterator 
of up->sg sg list[1]. This was nacked by you(?) [2] but I have problem 
understanding why? I see lot of code in kernel mapping sg pages:
linux$ git grep ' kmap.*sg' | wc -l
61
Is it incorrect? Or our case is different?

4. As you suggested in [3](?), modify xe_hmm_userptr_populate_range to 
keep hmm_range.hmm_pfns(or sth similar) in xe_userptr and use it later 
(instead of up->sg) to iterate over pages.

[1]: 
https://lore.kernel.org/intel-xe/20241011-xe_res_cursor_add_page_iterator-v3-1-0f8b8d3ab021@intel.com/
[2]: 
https://lore.kernel.org/intel-xe/Zw32fauoUmB6Iojk@DUT025-TGLU.fm.intel.com/
[3]: 
https://patchwork.freedesktop.org/patch/617481/?series=136572&rev=2#comment_1126527

Regards
Andrzej

> 
> Matt
> 
>> +	struct xe_vm *vm = xe_vma_vm(&uvma->vma);
>> +	struct xe_userptr *up = &uvma->userptr;
>> +	struct xe_res_cursor cur = {};
>> +	int cur_len, ret = 0;
>> +
>> +	while (true) {
>> +		down_read(&vm->userptr.notifier_lock);
>> +		if (!xe_vma_userptr_check_repin(uvma))
>> +			break;
>> +
>> +		spin_lock(&vm->userptr.invalidated_lock);
>> +		list_del_init(&uvma->userptr.invalidate_link);
>> +		spin_unlock(&vm->userptr.invalidated_lock);
>> +
>> +		up_read(&vm->userptr.notifier_lock);
>> +		ret = xe_vma_userptr_pin_pages(uvma);
>> +		if (ret)
>> +			return ret;
>> +	}
>> +
>> +	if (!up->sg) {
>> +		ret = -EINVAL;
>> +		goto out_unlock_notifier;
>> +	}
>> +
>> +	for (xe_res_first_sg(up->sg, offset, len, &cur); cur.remaining;
>> +	     xe_res_next(&cur, cur_len)) {
>> +		void *ptr = kmap_local_page(sg_page(cur.sgl)) + cur.start;
>> +
>> +		cur_len = min(cur.size, cur.remaining);
>> +		if (write)
>> +			memcpy(ptr, buf, cur_len);
>> +		else
>> +			memcpy(buf, ptr, cur_len);
>> +		kunmap_local(ptr);
>> +		buf += cur_len;
>> +	}
>> +	ret = len;
>> +
>> +out_unlock_notifier:
>> +	up_read(&vm->userptr.notifier_lock);
>> +	return ret;
>> +}
>> diff --git a/drivers/gpu/drm/xe/xe_vm.h b/drivers/gpu/drm/xe/xe_vm.h
>> index c864dba35e1d..99b9a9b011de 100644
>> --- a/drivers/gpu/drm/xe/xe_vm.h
>> +++ b/drivers/gpu/drm/xe/xe_vm.h
>> @@ -281,3 +281,6 @@ struct xe_vm_snapshot *xe_vm_snapshot_capture(struct xe_vm *vm);
>>   void xe_vm_snapshot_capture_delayed(struct xe_vm_snapshot *snap);
>>   void xe_vm_snapshot_print(struct xe_vm_snapshot *snap, struct drm_printer *p);
>>   void xe_vm_snapshot_free(struct xe_vm_snapshot *snap);
>> +
>> +int xe_uvma_access(struct xe_userptr_vma *uvma, u64 offset,
>> +		   void *buf, u64 len, bool write);
>> -- 
>> 2.34.1
>>



More information about the Intel-xe mailing list