[PATCH] drm/amdkfd: Have kfd driver use same PASID values from graphic driver

Felix Kuehling felix.kuehling at amd.com
Wed Nov 20 22:39:50 UTC 2024



On 2024-11-11 13:33, Xiaogang.Chen wrote:
> From: Xiaogang Chen <xiaogang.chen at amd.com>
> 
> Current kfd driver has its own PASID value for a kfd process and uses it to
> locate vm at interrupt handler or mapping between kfd process and vm. That
> design is not working when a physical gpu device has multiple spatial
> partitions, ex: adev in CPX mode. This patch has kfd driver use same pasid
> values that graphic driver generated which is per vm per pasid.
> 
> These pasid values are passed to fw/hardware. We do not need change interrupt
> handler though more pasid values are used. Also, pasid values at log are
> replaced by user process pid, or pasid values are not exposed to user, user
> will see their process pids that have meaning in user space.
> 
> Signed-off-by: Xiaogang Chen <xiaogang.chen at amd.com>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  4 +-
>  .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 34 +++++----
>  drivers/gpu/drm/amd/amdkfd/kfd_chardev.c      | 25 +++----
>  drivers/gpu/drm/amd/amdkfd/kfd_debug.c        |  8 +--
>  .../drm/amd/amdkfd/kfd_device_queue_manager.c | 52 +++++++-------
>  drivers/gpu/drm/amd/amdkfd/kfd_events.c       | 17 ++---
>  .../drm/amd/amdkfd/kfd_packet_manager_v9.c    |  4 +-
>  .../drm/amd/amdkfd/kfd_packet_manager_vi.c    |  3 +-
>  drivers/gpu/drm/amd/amdkfd/kfd_priv.h         |  5 +-
>  drivers/gpu/drm/amd/amdkfd/kfd_process.c      | 71 ++++++++-----------
>  .../amd/amdkfd/kfd_process_queue_manager.c    | 10 +--
>  drivers/gpu/drm/amd/amdkfd/kfd_svm.c          | 17 +++--
>  12 files changed, 127 insertions(+), 123 deletions(-)
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
> index 4b80ad860639..e576f65979a2 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
> @@ -47,6 +47,7 @@ enum TLB_FLUSH_TYPE {
>  };
>  
>  struct amdgpu_device;
> +struct kfd_process_device;
>  struct amdgpu_reset_context;
>  
>  enum kfd_mem_attachment_type {
> @@ -300,7 +301,8 @@ bool amdgpu_amdkfd_compute_active(struct amdgpu_device *adev, uint32_t node_id);
>  		((struct drm_file *)(drm_priv))->driver_priv)->vm)
>  
>  int amdgpu_amdkfd_gpuvm_set_vm_pasid(struct amdgpu_device *adev,
> -				     struct amdgpu_vm *avm, u32 pasid);
> +				     struct kfd_process_device *pdd,
> +				     struct amdgpu_vm *avm);
>  int amdgpu_amdkfd_gpuvm_acquire_process_vm(struct amdgpu_device *adev,
>  					struct amdgpu_vm *avm,
>  					void **process_info,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
> index f30548f4c3b3..f63c6b185bbb 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
> @@ -1530,24 +1530,30 @@ static void amdgpu_amdkfd_gpuvm_unpin_bo(struct amdgpu_bo *bo)
>  }
>  
>  int amdgpu_amdkfd_gpuvm_set_vm_pasid(struct amdgpu_device *adev,
> -				     struct amdgpu_vm *avm, u32 pasid)
> +				     struct kfd_process_device *pdd,
> +				     struct amdgpu_vm *avm)

I'd prefer if this function got renamed to amdgpu_amdkfd_gpuvm_get_pasid and it returns the pasid.

>  
>  {
> -	int ret;
> -
> -	/* Free the original amdgpu allocated pasid,
> -	 * will be replaced with kfd allocated pasid.
> -	 */
> -	if (avm->pasid) {
> -		amdgpu_pasid_free(avm->pasid);
> -		amdgpu_vm_set_pasid(adev, avm, 0);
> +	int ret = 0;
> +	int pasid;
> +
> +	/* if avm has pasid assigned use it */
> +	if (avm->pasid)
> +		pdd->pasid = avm->pasid;
> +	else {
> +		/* otherwise generate a new pasid and assign to avm */

Does this ever happen? I think any avm that was acquired properly should already have a PASID. If it doesn't, maybe we can just return an error.

> +		pasid = amdgpu_pasid_alloc(16);
> +		if (pasid < 0) {
> +			dev_warn(adev->dev, "No more PASIDs available!");
> +			ret = -ENOMEM;
> +		} else
> +			ret = xa_err(xa_store_irq(&adev->vm_manager.pasids, pasid, avm,
> +						  GFP_KERNEL));
> +		if (ret >= 0)
> +			avm->pasid = pasid;
>  	}
>  
> -	ret = amdgpu_vm_set_pasid(adev, avm, pasid);
> -	if (ret)
> -		return ret;
> -
> -	return 0;
> +	return ret;
>  }
>  
>  int amdgpu_amdkfd_gpuvm_acquire_process_vm(struct amdgpu_device *adev,
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> index 065d87841459..8b1fd8e7452b 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> @@ -155,8 +155,8 @@ static int kfd_open(struct inode *inode, struct file *filep)
>  	/* filep now owns the reference returned by kfd_create_process */
>  	filep->private_data = process;
>  
> -	dev_dbg(kfd_device, "process %d opened, compat mode (32 bit) - %d\n",
> -		process->pasid, process->is_32bit_user_mode);
> +	dev_dbg(kfd_device, "process pid %d opened kfd node, compat mode (32 bit) - %d\n",
> +		(int)process->lead_thread->pid, process->is_32bit_user_mode);
>  
>  	return 0;
>  }
> @@ -361,8 +361,8 @@ static int kfd_ioctl_create_queue(struct file *filep, struct kfd_process *p,
>  		goto err_acquire_queue_buf;
>  	}
>  
> -	pr_debug("Creating queue for PASID 0x%x on gpu 0x%x\n",
> -			p->pasid,
> +	pr_debug("Creating queue for process pid 0x%x on gpu 0x%x\n",
> +			(int)p->lead_thread->pid,
>  			dev->id);
>  
>  	err = pqm_create_queue(&p->pqm, dev, &q_properties, &queue_id,
> @@ -415,9 +415,9 @@ static int kfd_ioctl_destroy_queue(struct file *filp, struct kfd_process *p,
>  	int retval;
>  	struct kfd_ioctl_destroy_queue_args *args = data;
>  
> -	pr_debug("Destroying queue id %d for pasid 0x%x\n",
> +	pr_debug("Destroying queue id %d for process pid 0x%x\n",
>  				args->queue_id,
> -				p->pasid);
> +				(int)p->lead_thread->pid);
>  
>  	mutex_lock(&p->mutex);
>  
> @@ -468,8 +468,8 @@ static int kfd_ioctl_update_queue(struct file *filp, struct kfd_process *p,
>  	properties.pm4_target_xcc = (args->queue_percentage >> 8) & 0xFF;
>  	properties.priority = args->queue_priority;
>  
> -	pr_debug("Updating queue id %d for pasid 0x%x\n",
> -			args->queue_id, p->pasid);
> +	pr_debug("Updating queue id %d for process pid 0x%x\n",
> +			args->queue_id, (int)p->lead_thread->pid);
>  
>  	mutex_lock(&p->mutex);
>  
> @@ -695,7 +695,7 @@ static int kfd_ioctl_get_process_apertures(struct file *filp,
>  	struct kfd_process_device_apertures *pAperture;
>  	int i;
>  
> -	dev_dbg(kfd_device, "get apertures for PASID 0x%x", p->pasid);
> +	dev_dbg(kfd_device, "get apertures for process pid 0x%x", (int)p->lead_thread->pid);
>  
>  	args->num_of_nodes = 0;
>  
> @@ -747,7 +747,8 @@ static int kfd_ioctl_get_process_apertures_new(struct file *filp,
>  	int ret;
>  	int i;
>  
> -	dev_dbg(kfd_device, "get apertures for PASID 0x%x", p->pasid);
> +	dev_dbg(kfd_device, "get apertures for process pid 0x%x",
> +			(int)p->lead_thread->pid);
>  
>  	if (args->num_of_nodes == 0) {
>  		/* Return number of nodes, so that user space can alloacate
> @@ -3365,12 +3366,12 @@ static int kfd_mmio_mmap(struct kfd_node *dev, struct kfd_process *process,
>  
>  	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
>  
> -	pr_debug("pasid 0x%x mapping mmio page\n"
> +	pr_debug("process pid 0x%x mapping mmio page\n"
>  		 "     target user address == 0x%08llX\n"
>  		 "     physical address    == 0x%08llX\n"
>  		 "     vm_flags            == 0x%04lX\n"
>  		 "     size                == 0x%04lX\n",
> -		 process->pasid, (unsigned long long) vma->vm_start,
> +		 (int)process->lead_thread->pid, (unsigned long long) vma->vm_start,
>  		 address, vma->vm_flags, PAGE_SIZE);
>  
>  	return io_remap_pfn_range(vma,
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_debug.c b/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> index 312dfa84f29f..93ba497042b2 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> @@ -238,8 +238,8 @@ bool kfd_set_dbg_ev_from_interrupt(struct kfd_node *dev,
>  
>  			mutex_unlock(&p->mutex);
>  		} else if (trap_mask & KFD_EC_MASK(EC_DEVICE_MEMORY_VIOLATION)) {
> -			kfd_dqm_evict_pasid(dev->dqm, p->pasid);
> -			kfd_signal_vm_fault_event(dev, p->pasid, NULL,
> +			kfd_dqm_evict_pasid(dev->dqm, pasid);
> +			kfd_signal_vm_fault_event(dev, pasid, NULL,
>  							exception_data);
>  
>  			signaled_to_debugger_or_runtime = true;
> @@ -276,8 +276,8 @@ int kfd_dbg_send_exception_to_runtime(struct kfd_process *p,
>  		data = (struct kfd_hsa_memory_exception_data *)
>  						pdd->vm_fault_exc_data;
>  
> -		kfd_dqm_evict_pasid(pdd->dev->dqm, p->pasid);
> -		kfd_signal_vm_fault_event(pdd->dev, p->pasid, NULL, data);
> +		kfd_dqm_evict_pasid(pdd->dev->dqm, pdd->pasid);

kfd_dqm_evict_pasid just goes and looks up the pdd from the pasid again. Maybe this could be streamlined with a kfd_dqm_evict_pdd that avoids the unnecessary lookup. Could be a follow-up patch.


> +		kfd_signal_vm_fault_event(pdd->dev, pdd->pasid, NULL, data);

Similarly, this function looks up the process from the PASID. Maybe this could be refactored to avoid the redundant lookup.

>  		error_reason &= ~KFD_EC_MASK(EC_DEVICE_MEMORY_VIOLATION);
>  	}
>  
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> index 38c19dc8311d..ca8db5e3a7b7 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> @@ -206,7 +206,7 @@ static int add_queue_mes(struct device_queue_manager *dqm, struct queue *q,
>  		return -EIO;
>  
>  	memset(&queue_input, 0x0, sizeof(struct mes_add_queue_input));
> -	queue_input.process_id = qpd->pqm->process->pasid;
> +	queue_input.process_id = pdd->pasid;
>  	queue_input.page_table_base_addr =  qpd->page_table_base;
>  	queue_input.process_va_start = 0;
>  	queue_input.process_va_end = adev->vm_manager.max_pfn - 1;
> @@ -496,6 +496,7 @@ static int allocate_vmid(struct device_queue_manager *dqm,
>  			struct qcm_process_device *qpd,
>  			struct queue *q)
>  {
> +	struct kfd_process_device *pdd = qpd_to_pdd(qpd);
>  	struct device *dev = dqm->dev->adev->dev;
>  	int allocated_vmid = -1, i;
>  
> @@ -514,9 +515,9 @@ static int allocate_vmid(struct device_queue_manager *dqm,
>  
>  	pr_debug("vmid allocated: %d\n", allocated_vmid);
>  
> -	dqm->vmid_pasid[allocated_vmid] = q->process->pasid;
> +	dqm->vmid_pasid[allocated_vmid] = pdd->pasid;
>  
> -	set_pasid_vmid_mapping(dqm, q->process->pasid, allocated_vmid);
> +	set_pasid_vmid_mapping(dqm, pdd->pasid, allocated_vmid);
>  
>  	qpd->vmid = allocated_vmid;
>  	q->properties.vmid = allocated_vmid;
> @@ -768,6 +769,11 @@ static int dbgdev_wave_reset_wavefronts(struct kfd_node *dev, struct kfd_process
>  		return -EOPNOTSUPP;
>  	}
>  
> +	/* taking the VMID for that process on the safe way using PDD */
> +	pdd = kfd_get_process_device_data(dev, p);
> +	if (!pdd)
> +		return -EFAULT;
> +
>  	/* Scan all registers in the range ATC_VMID8_PASID_MAPPING ..
>  	 * ATC_VMID15_PASID_MAPPING
>  	 * to check which VMID the current process is mapped to.
> @@ -777,23 +783,19 @@ static int dbgdev_wave_reset_wavefronts(struct kfd_node *dev, struct kfd_process
>  		status = dev->kfd2kgd->get_atc_vmid_pasid_mapping_info
>  				(dev->adev, vmid, &queried_pasid);
>  
> -		if (status && queried_pasid == p->pasid) {
> -			pr_debug("Killing wave fronts of vmid %d and pasid 0x%x\n",
> -					vmid, p->pasid);
> +		if (status && queried_pasid == pdd->pasid) {
> +			pr_debug("Killing wave fronts of vmid %d and process pid 0x%x\n",
> +				 vmid, p->lead_thread->pid);
>  			break;
>  		}
>  	}
>  
>  	if (vmid > last_vmid_to_scan) {
> -		dev_err(dev->adev->dev, "Didn't find vmid for pasid 0x%x\n", p->pasid);
> +		dev_err(dev->adev->dev, "Didn't find vmid for process pid 0x%x\n",
> +			p->lead_thread->pid);
>  		return -EFAULT;
>  	}
>  
> -	/* taking the VMID for that process on the safe way using PDD */
> -	pdd = kfd_get_process_device_data(dev, p);
> -	if (!pdd)
> -		return -EFAULT;
> -
>  	reg_gfx_index.bits.sh_broadcast_writes = 1;
>  	reg_gfx_index.bits.se_broadcast_writes = 1;
>  	reg_gfx_index.bits.instance_broadcast_writes = 1;
> @@ -1029,8 +1031,8 @@ static int suspend_single_queue(struct device_queue_manager *dqm,
>  	if (q->properties.is_suspended)
>  		return 0;
>  
> -	pr_debug("Suspending PASID %u queue [%i]\n",
> -			pdd->process->pasid,
> +	pr_debug("Suspending process pid %u queue [%i]\n",
> +			pdd->process->lead_thread->pid,
>  			q->properties.queue_id);
>  
>  	is_new = q->properties.exception_status & KFD_EC_MASK(EC_QUEUE_NEW);
> @@ -1077,8 +1079,8 @@ static int resume_single_queue(struct device_queue_manager *dqm,
>  
>  	pdd = qpd_to_pdd(qpd);
>  
> -	pr_debug("Restoring from suspend PASID %u queue [%i]\n",
> -			    pdd->process->pasid,
> +	pr_debug("Restoring from suspend process pid %u queue [%i]\n",
> +			    pdd->process->lead_thread->pid,
>  			    q->properties.queue_id);
>  
>  	q->properties.is_suspended = false;
> @@ -1111,8 +1113,8 @@ static int evict_process_queues_nocpsch(struct device_queue_manager *dqm,
>  		goto out;
>  
>  	pdd = qpd_to_pdd(qpd);
> -	pr_debug_ratelimited("Evicting PASID 0x%x queues\n",
> -			    pdd->process->pasid);
> +	pr_debug_ratelimited("Evicting process pid 0x%x queues\n",
> +			    pdd->process->lead_thread->pid);
>  
>  	pdd->last_evict_timestamp = get_jiffies_64();
>  	/* Mark all queues as evicted. Deactivate all active queues on
> @@ -1169,8 +1171,8 @@ static int evict_process_queues_cpsch(struct device_queue_manager *dqm,
>  	if (!pdd->drm_priv)
>  		goto out;
>  
> -	pr_debug_ratelimited("Evicting PASID 0x%x queues\n",
> -			    pdd->process->pasid);
> +	pr_debug_ratelimited("Evicting process pid 0x%x queues\n",
> +			    pdd->process->lead_thread->pid);
>  
>  	/* Mark all queues as evicted. Deactivate all active queues on
>  	 * the qpd.
> @@ -1228,8 +1230,8 @@ static int restore_process_queues_nocpsch(struct device_queue_manager *dqm,
>  		goto out;
>  	}
>  
> -	pr_debug_ratelimited("Restoring PASID 0x%x queues\n",
> -			    pdd->process->pasid);
> +	pr_debug_ratelimited("Restoring process pid 0x%x queues\n",
> +			    pdd->process->lead_thread->pid);
>  
>  	/* Update PD Base in QPD */
>  	qpd->page_table_base = pd_base;
> @@ -1312,8 +1314,8 @@ static int restore_process_queues_cpsch(struct device_queue_manager *dqm,
>  	if (!pdd->drm_priv)
>  		goto vm_not_acquired;
>  
> -	pr_debug_ratelimited("Restoring PASID 0x%x queues\n",
> -			    pdd->process->pasid);
> +	pr_debug_ratelimited("Restoring process pid 0x%x queues\n",
> +			     pdd->process->lead_thread->pid);
>  
>  	/* Update PD Base in QPD */
>  	qpd->page_table_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv);
> @@ -2102,7 +2104,7 @@ static void set_queue_as_reset(struct device_queue_manager *dqm, struct queue *q
>  	struct kfd_process_device *pdd = qpd_to_pdd(qpd);
>  
>  	dev_err(dqm->dev->adev->dev, "queue id 0x%0x at pasid 0x%0x is reset\n",
> -		q->properties.queue_id, q->process->pasid);
> +		q->properties.queue_id, pdd->process->lead_thread->pid);
>  
>  	pdd->has_reset_queue = true;
>  	if (q->properties.is_active) {
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_events.c b/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> index ea3792249209..a38c0b3b88fc 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> @@ -1128,8 +1128,8 @@ static void lookup_events_by_type_and_signal(struct kfd_process *p,
>  
>  	if (type == KFD_EVENT_TYPE_MEMORY) {
>  		dev_warn(kfd_device,
> -			"Sending SIGSEGV to process %d (pasid 0x%x)",
> -				p->lead_thread->pid, p->pasid);
> +			"Sending SIGSEGV to process pid %d",
> +				p->lead_thread->pid);
>  		send_sig(SIGSEGV, p->lead_thread, 0);
>  	}
>  
> @@ -1137,13 +1137,13 @@ static void lookup_events_by_type_and_signal(struct kfd_process *p,
>  	if (send_signal) {
>  		if (send_sigterm) {
>  			dev_warn(kfd_device,
> -				"Sending SIGTERM to process %d (pasid 0x%x)",
> -					p->lead_thread->pid, p->pasid);
> +				"Sending SIGTERM to process pid %d",
> +					p->lead_thread->pid);
>  			send_sig(SIGTERM, p->lead_thread, 0);
>  		} else {
>  			dev_err(kfd_device,
> -				"Process %d (pasid 0x%x) got unhandled exception",
> -				p->lead_thread->pid, p->pasid);
> +				"Process pid %d got unhandled exception",
> +				p->lead_thread->pid);
>  		}
>  	}
>  
> @@ -1253,7 +1253,8 @@ void kfd_signal_reset_event(struct kfd_node *dev)
>  		}
>  
>  		if (unlikely(!pdd)) {
> -			WARN_ONCE(1, "Could not get device data from pasid:0x%x\n", p->pasid);
> +			WARN_ONCE(1, "Could not get device data from process pid:0x%x\n",
> +				  p->lead_thread->pid);
>  			continue;
>  		}
>  
> @@ -1263,7 +1264,7 @@ void kfd_signal_reset_event(struct kfd_node *dev)
>  		if (dev->dqm->detect_hang_count) {
>  			struct amdgpu_task_info *ti;
>  
> -			ti = amdgpu_vm_get_task_info_pasid(dev->adev, p->pasid);
> +			ti = amdgpu_vm_get_task_info_pasid(dev->adev, pdd->pasid);

Can we use amdgpu_vm_get_task_info_vm here to avoid a redundant lookup of the VM from the pasid? You'd need a helper to get the vm from pdd->drm_priv (drm_priv_to_vm in amdgpu_amdkfd.h).

>  			if (ti) {
>  				dev_err(dev->adev->dev,
>  					"Queues reset on process %s tid %d thread %s pid %d\n",
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_v9.c b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_v9.c
> index 1f9f5bfeaf86..d56525201155 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_v9.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_v9.c
> @@ -47,7 +47,7 @@ static int pm_map_process_v9(struct packet_manager *pm,
>  		packet->bitfields2.exec_cleaner_shader = 1;
>  	packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0;
>  	packet->bitfields2.process_quantum = 10;
> -	packet->bitfields2.pasid = qpd->pqm->process->pasid;
> +	packet->bitfields2.pasid = pdd->pasid;
>  	packet->bitfields14.gds_size = qpd->gds_size & 0x3F;
>  	packet->bitfields14.gds_size_hi = (qpd->gds_size >> 6) & 0xF;
>  	packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd->num_gws : 0;
> @@ -106,7 +106,7 @@ static int pm_map_process_aldebaran(struct packet_manager *pm,
>  		packet->bitfields2.exec_cleaner_shader = 1;
>  	packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0;
>  	packet->bitfields2.process_quantum = 10;
> -	packet->bitfields2.pasid = qpd->pqm->process->pasid;
> +	packet->bitfields2.pasid = pdd->pasid;
>  	packet->bitfields14.gds_size = qpd->gds_size & 0x3F;
>  	packet->bitfields14.gds_size_hi = (qpd->gds_size >> 6) & 0xF;
>  	packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd->num_gws : 0;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_vi.c b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_vi.c
> index c1199d06d131..347c86e1c378 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_vi.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_vi.c
> @@ -42,6 +42,7 @@ unsigned int pm_build_pm4_header(unsigned int opcode, size_t packet_size)
>  static int pm_map_process_vi(struct packet_manager *pm, uint32_t *buffer,
>  				struct qcm_process_device *qpd)
>  {
> +	struct kfd_process_device *pdd = qpd_to_pdd(qpd);
>  	struct pm4_mes_map_process *packet;
>  
>  	packet = (struct pm4_mes_map_process *)buffer;
> @@ -52,7 +53,7 @@ static int pm_map_process_vi(struct packet_manager *pm, uint32_t *buffer,
>  					sizeof(struct pm4_mes_map_process));
>  	packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0;
>  	packet->bitfields2.process_quantum = 10;
> -	packet->bitfields2.pasid = qpd->pqm->process->pasid;
> +	packet->bitfields2.pasid = pdd->pasid;
>  	packet->bitfields3.page_table_base = qpd->page_table_base;
>  	packet->bitfields10.gds_size = qpd->gds_size;
>  	packet->bitfields10.num_gws = qpd->num_gws;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> index 9e5ca0b93b2a..08bcbc45197c 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> @@ -849,6 +849,8 @@ struct kfd_process_device {
>  
>  	/* Tracks queue reset status */
>  	bool has_reset_queue;
> +
> +	u32 pasid;
>  };
>  
>  #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd)
> @@ -908,8 +910,6 @@ struct kfd_process {
>  	/* We want to receive a notification when the mm_struct is destroyed */
>  	struct mmu_notifier mmu_notifier;
>  
> -	u32 pasid;
> -
>  	/*
>  	 * Array of kfd_process_device pointers,
>  	 * one for each device the process is using.
> @@ -970,7 +970,6 @@ struct kfd_process {
>  	/* Kobj for our procfs */
>  	struct kobject *kobj;
>  	struct kobject *kobj_queues;
> -	struct attribute attr_pasid;
>  
>  	/* Keep track cwsr init */
>  	bool has_cwsr;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process.c b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> index 6bab6fc6a35d..edc4f47061f5 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> @@ -282,8 +282,8 @@ static int kfd_get_cu_occupancy(struct attribute *attr, char *buffer)
>  	cu_cnt = 0;
>  	proc = pdd->process;
>  	if (pdd->qpd.queue_count == 0) {
> -		pr_debug("Gpu-Id: %d has no active queues for process %d\n",
> -			 dev->id, proc->pasid);
> +		pr_debug("Gpu-Id: %d has no active queues for process pid %d\n",
> +			 dev->id, (int)proc->lead_thread->pid);
>  		return snprintf(buffer, PAGE_SIZE, "%d\n", cu_cnt);
>  	}
>  
> @@ -327,12 +327,7 @@ static int kfd_get_cu_occupancy(struct attribute *attr, char *buffer)
>  static ssize_t kfd_procfs_show(struct kobject *kobj, struct attribute *attr,
>  			       char *buffer)
>  {
> -	if (strcmp(attr->name, "pasid") == 0) {
> -		struct kfd_process *p = container_of(attr, struct kfd_process,
> -						     attr_pasid);
> -
> -		return snprintf(buffer, PAGE_SIZE, "%d\n", p->pasid);
> -	} else if (strncmp(attr->name, "vram_", 5) == 0) {
> +	if (strncmp(attr->name, "vram_", 5) == 0) {
>  		struct kfd_process_device *pdd = container_of(attr, struct kfd_process_device,
>  							      attr_vram);
>  		return snprintf(buffer, PAGE_SIZE, "%llu\n", atomic64_read(&pdd->vram_usage));
> @@ -887,9 +882,6 @@ struct kfd_process *kfd_create_process(struct task_struct *thread)
>  			goto out;
>  		}
>  
> -		kfd_sysfs_create_file(process->kobj, &process->attr_pasid,
> -				      "pasid");
> -
>  		process->kobj_queues = kobject_create_and_add("queues",
>  							process->kobj);
>  		if (!process->kobj_queues)
> @@ -1055,8 +1047,8 @@ static void kfd_process_destroy_pdds(struct kfd_process *p)
>  	for (i = 0; i < p->n_pdds; i++) {
>  		struct kfd_process_device *pdd = p->pdds[i];
>  
> -		pr_debug("Releasing pdd (topology id %d) for process (pasid 0x%x)\n",
> -				pdd->dev->id, p->pasid);
> +		pr_debug("Releasing pdd (topology id %d, for pid 0x%0x)\n",
> +			 pdd->dev->id, (int)p->lead_thread->pid);
>  
>  		kfd_process_device_destroy_cwsr_dgpu(pdd);
>  		kfd_process_device_destroy_ib_mem(pdd);
> @@ -1102,7 +1094,6 @@ static void kfd_process_remove_sysfs(struct kfd_process *p)
>  	if (!p->kobj)
>  		return;
>  
> -	sysfs_remove_file(p->kobj, &p->attr_pasid);
>  	kobject_del(p->kobj_queues);
>  	kobject_put(p->kobj_queues);
>  	p->kobj_queues = NULL;
> @@ -1171,7 +1162,6 @@ static void kfd_process_wq_release(struct work_struct *work)
>  
>  	kfd_event_free_process(p);
>  
> -	kfd_pasid_free(p->pasid);
>  	mutex_destroy(&p->mutex);
>  
>  	put_task_struct(p->lead_thread);
> @@ -1524,12 +1514,6 @@ static struct kfd_process *create_process(const struct task_struct *thread)
>  	atomic_set(&process->debugged_process_count, 0);
>  	sema_init(&process->runtime_enable_sema, 0);
>  
> -	process->pasid = kfd_pasid_alloc();
> -	if (process->pasid == 0) {
> -		err = -ENOSPC;
> -		goto err_alloc_pasid;
> -	}
> -
>  	err = pqm_init(&process->pqm, process);
>  	if (err != 0)
>  		goto err_process_pqm_init;
> @@ -1583,8 +1567,6 @@ static struct kfd_process *create_process(const struct task_struct *thread)
>  err_init_apertures:
>  	pqm_uninit(&process->pqm);
>  err_process_pqm_init:
> -	kfd_pasid_free(process->pasid);
> -err_alloc_pasid:
>  	kfd_event_free_process(process);
>  err_event_init:
>  	mutex_destroy(&process->mutex);
> @@ -1723,7 +1705,7 @@ int kfd_process_device_init_vm(struct kfd_process_device *pdd,
>  	if (ret)
>  		goto err_init_cwsr;
>  
> -	ret = amdgpu_amdkfd_gpuvm_set_vm_pasid(dev->adev, avm, p->pasid);
> +	ret = amdgpu_amdkfd_gpuvm_set_vm_pasid(dev->adev, pdd, avm);
>  	if (ret)
>  		goto err_set_pasid;
>  
> @@ -1822,15 +1804,20 @@ struct kfd_process *kfd_lookup_process_by_pasid(u32 pasid)
>  {
>  	struct kfd_process *p, *ret_p = NULL;
>  	unsigned int temp;
> +	int i;
>  
>  	int idx = srcu_read_lock(&kfd_processes_srcu);
>  
>  	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
> -		if (p->pasid == pasid) {
> -			kref_get(&p->ref);
> -			ret_p = p;
> -			break;
> +		for (i = 0; i < p->n_pdds; i++) {
> +			if (p->pdds[i]->pasid == pasid) {
> +				kref_get(&p->ref);
> +				ret_p = p;
> +				break;
> +			}

I think this won't work correctly. The same PASID can be used for different processes on different GPUs because each adev manages its own PASID->amdgpu_vm lookup table. So kfd_lookup_process_by_pasid needs a new parameter that identifies the GPU adev, and you should only compare pasids, if the adev matches.

Regards,
  Felix

>  		}
> +		if (ret_p)
> +			break;
>  	}
>  
>  	srcu_read_unlock(&kfd_processes_srcu, idx);
> @@ -1991,7 +1978,7 @@ static void evict_process_worker(struct work_struct *work)
>  	 */
>  	p = container_of(dwork, struct kfd_process, eviction_work);
>  
> -	pr_debug("Started evicting pasid 0x%x\n", p->pasid);
> +	pr_debug("Started evicting process pid 0x%x\n", (int)p->lead_thread->pid);
>  	ret = kfd_process_evict_queues(p, KFD_QUEUE_EVICTION_TRIGGER_TTM);
>  	if (!ret) {
>  		/* If another thread already signaled the eviction fence,
> @@ -2003,9 +1990,9 @@ static void evict_process_worker(struct work_struct *work)
>  				     msecs_to_jiffies(PROCESS_RESTORE_TIME_MS)))
>  			kfd_process_restore_queues(p);
>  
> -		pr_debug("Finished evicting pasid 0x%x\n", p->pasid);
> +		pr_debug("Finished evicting process pid 0x%x\n", (int)p->lead_thread->pid);
>  	} else
> -		pr_err("Failed to evict queues of pasid 0x%x\n", p->pasid);
> +		pr_err("Failed to evict queues of process pid 0x%x\n", (int)p->lead_thread->pid);
>  }
>  
>  static int restore_process_helper(struct kfd_process *p)
> @@ -2022,9 +2009,11 @@ static int restore_process_helper(struct kfd_process *p)
>  
>  	ret = kfd_process_restore_queues(p);
>  	if (!ret)
> -		pr_debug("Finished restoring pasid 0x%x\n", p->pasid);
> +		pr_debug("Finished restoring process pid 0x%x\n",
> +			 (int)p->lead_thread->pid);
>  	else
> -		pr_err("Failed to restore queues of pasid 0x%x\n", p->pasid);
> +		pr_err("Failed to restore queues of process pid 0x%x\n",
> +			(int)p->lead_thread->pid);
>  
>  	return ret;
>  }
> @@ -2041,7 +2030,7 @@ static void restore_process_worker(struct work_struct *work)
>  	 * lifetime of this thread, kfd_process p will be valid
>  	 */
>  	p = container_of(dwork, struct kfd_process, restore_work);
> -	pr_debug("Started restoring pasid 0x%x\n", p->pasid);
> +	pr_debug("Started restoring process pasid 0x%x\n", (int)p->lead_thread->pid);
>  
>  	/* Setting last_restore_timestamp before successful restoration.
>  	 * Otherwise this would have to be set by KGD (restore_process_bos)
> @@ -2057,8 +2046,8 @@ static void restore_process_worker(struct work_struct *work)
>  
>  	ret = restore_process_helper(p);
>  	if (ret) {
> -		pr_debug("Failed to restore BOs of pasid 0x%x, retry after %d ms\n",
> -			 p->pasid, PROCESS_BACK_OFF_TIME_MS);
> +		pr_debug("Failed to restore BOs of process pid 0x%x, retry after %d ms\n",
> +			 (int)p->lead_thread->pid, PROCESS_BACK_OFF_TIME_MS);
>  		if (mod_delayed_work(kfd_restore_wq, &p->restore_work,
>  				     msecs_to_jiffies(PROCESS_RESTORE_TIME_MS)))
>  			kfd_process_restore_queues(p);
> @@ -2074,7 +2063,7 @@ void kfd_suspend_all_processes(void)
>  	WARN(debug_evictions, "Evicting all processes");
>  	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
>  		if (kfd_process_evict_queues(p, KFD_QUEUE_EVICTION_TRIGGER_SUSPEND))
> -			pr_err("Failed to suspend process 0x%x\n", p->pasid);
> +			pr_err("Failed to suspend process pid 0x%x\n", (int)p->lead_thread->pid);
>  		signal_eviction_fence(p);
>  	}
>  	srcu_read_unlock(&kfd_processes_srcu, idx);
> @@ -2088,8 +2077,8 @@ int kfd_resume_all_processes(void)
>  
>  	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
>  		if (restore_process_helper(p)) {
> -			pr_err("Restore process %d failed during resume\n",
> -			       p->pasid);
> +			pr_err("Restore process pid %d failed during resume\n",
> +			       (int)p->lead_thread->pid);
>  			ret = -EFAULT;
>  		}
>  	}
> @@ -2144,7 +2133,7 @@ int kfd_process_drain_interrupts(struct kfd_process_device *pdd)
>  	memset(irq_drain_fence, 0, sizeof(irq_drain_fence));
>  	irq_drain_fence[0] = (KFD_IRQ_FENCE_SOURCEID << 8) |
>  							KFD_IRQ_FENCE_CLIENTID;
> -	irq_drain_fence[3] = pdd->process->pasid;
> +	irq_drain_fence[3] = pdd->pasid;
>  
>  	/*
>  	 * For GFX 9.4.3, send the NodeId also in IH cookie DW[3]
> @@ -2296,7 +2285,7 @@ int kfd_debugfs_mqds_by_process(struct seq_file *m, void *data)
>  
>  	hash_for_each_rcu(kfd_processes_table, temp, p, kfd_processes) {
>  		seq_printf(m, "Process %d PASID 0x%x:\n",
> -			   p->lead_thread->tgid, p->pasid);
> +			   p->lead_thread->tgid, (int)p->lead_thread->pid);
>  
>  		mutex_lock(&p->mutex);
>  		r = pqm_debugfs_mqds(m, &p->pqm);
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> index c76db22a1000..1f98c16e0fe9 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> @@ -69,8 +69,8 @@ static int find_available_queue_slot(struct process_queue_manager *pqm,
>  	pr_debug("The new slot id %lu\n", found);
>  
>  	if (found >= KFD_MAX_NUM_OF_QUEUES_PER_PROCESS) {
> -		pr_info("Cannot open more queues for process with pasid 0x%x\n",
> -				pqm->process->pasid);
> +		pr_info("Cannot open more queues for process with pid 0x%x\n",
> +			pqm->process->lead_thread->pid);
>  		return -ENOMEM;
>  	}
>  
> @@ -425,8 +425,8 @@ int pqm_create_queue(struct process_queue_manager *pqm,
>  	}
>  
>  	if (retval != 0) {
> -		pr_err("Pasid 0x%x DQM create queue type %d failed. ret %d\n",
> -			pqm->process->pasid, type, retval);
> +		pr_err("process pid 0x%x DQM create queue type %d failed. ret %d\n",
> +			pqm->process->lead_thread->pid, type, retval);
>  		goto err_create_queue;
>  	}
>  
> @@ -520,7 +520,7 @@ int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid)
>  		retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q);
>  		if (retval) {
>  			pr_err("Pasid 0x%x destroy queue %d failed, ret %d\n",
> -				pqm->process->pasid,
> +				pdd->pasid,
>  				pqn->q->properties.queue_id, retval);
>  			if (retval != -ETIME)
>  				goto err_destroy_queue;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> index 3e2911895c74..4910cfcaff36 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> @@ -563,7 +563,8 @@ svm_range_vram_node_new(struct kfd_node *node, struct svm_range *prange,
>  	int r;
>  
>  	p = container_of(prange->svms, struct kfd_process, svms);
> -	pr_debug("pasid: %x svms 0x%p [0x%lx 0x%lx]\n", p->pasid, prange->svms,
> +	pr_debug("process pid: %x svms 0x%p [0x%lx 0x%lx]\n",
> +		 (int)p->lead_thread->pid, prange->svms,
>  		 prange->start, prange->last);
>  
>  	if (svm_range_validate_svm_bo(node, prange))
> @@ -3232,7 +3233,8 @@ void svm_range_list_fini(struct kfd_process *p)
>  	struct svm_range *prange;
>  	struct svm_range *next;
>  
> -	pr_debug("pasid 0x%x svms 0x%p\n", p->pasid, &p->svms);
> +	pr_debug("process pid 0x%x svms 0x%p\n", (int)p->lead_thread->pid,
> +		 &p->svms);
>  
>  	cancel_delayed_work_sync(&p->svms.restore_work);
>  
> @@ -3255,7 +3257,8 @@ void svm_range_list_fini(struct kfd_process *p)
>  
>  	mutex_destroy(&p->svms.lock);
>  
> -	pr_debug("pasid 0x%x svms 0x%p done\n", p->pasid, &p->svms);
> +	pr_debug("process pid 0x%x svms 0x%p done\n",
> +		(int)p->lead_thread->pid, &p->svms);
>  }
>  
>  int svm_range_list_init(struct kfd_process *p)
> @@ -3618,8 +3621,8 @@ svm_range_set_attr(struct kfd_process *p, struct mm_struct *mm,
>  	bool flush_tlb;
>  	int r, ret = 0;
>  
> -	pr_debug("pasid 0x%x svms 0x%p [0x%llx 0x%llx] pages 0x%llx\n",
> -		 p->pasid, &p->svms, start, start + size - 1, size);
> +	pr_debug("process pid 0x%x svms 0x%p [0x%llx 0x%llx] pages 0x%llx\n",
> +		 (int)p->lead_thread->pid, &p->svms, start, start + size - 1, size);
>  
>  	r = svm_range_check_attr(p, nattr, attrs);
>  	if (r)
> @@ -3727,8 +3730,8 @@ svm_range_set_attr(struct kfd_process *p, struct mm_struct *mm,
>  out:
>  	mutex_unlock(&process_info->lock);
>  
> -	pr_debug("pasid 0x%x svms 0x%p [0x%llx 0x%llx] done, r=%d\n", p->pasid,
> -		 &p->svms, start, start + size - 1, r);
> +	pr_debug("process pid 0x%x svms 0x%p [0x%llx 0x%llx] done, r=%d\n",
> +		 (int)p->lead_thread->pid, &p->svms, start, start + size - 1, r);
>  
>  	return ret ? ret : r;
>  }


More information about the amd-gfx mailing list