[PATCH 06/19] drm/amdkfd: Consolidate and clean up log commands
Oded Gabbay
oded.gabbay at gmail.com
Sat Aug 12 13:01:40 UTC 2017
On Sat, Aug 12, 2017 at 12:56 AM, Felix Kuehling <Felix.Kuehling at amd.com> wrote:
> From: Kent Russell <kent.russell at amd.com>
>
> Consolidate log commands so that dev_info(NULL, "Error...") uses the more
> accurate pr_err, remove the module name from the log (can be seen via
> dynamic debugging with +m), and the function name (can be seen via
> dynamic debugging with +f). We also don't need debug messages saying
> what function we're in. Those can be added by devs when needed
>
> Don't print vendor and device ID in error messages. They are typically
> the same for all GPUs in a multi-GPU system. So this doesn't add any
> value to the message.
>
> Lastly, remove parentheses around %d, %i and 0x%llX.
> According to kernel.org:
> "Printing numbers in parentheses (%d) adds no value and should be
> avoided."
>
> Signed-off-by: Kent Russell <kent.russell at amd.com>
> Signed-off-by: Yong Zhao <Yong.Zhao at amd.com>
> Signed-off-by: Felix Kuehling <Felix.Kuehling at amd.com>
> ---
> drivers/gpu/drm/amd/amdkfd/kfd_chardev.c | 64 ++++++++---------
> drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c | 38 +++++-----
> drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.c | 4 +-
> drivers/gpu/drm/amd/amdkfd/kfd_device.c | 51 ++++++--------
> .../gpu/drm/amd/amdkfd/kfd_device_queue_manager.c | 81 +++++++---------------
> .../drm/amd/amdkfd/kfd_device_queue_manager_cik.c | 2 +-
> .../drm/amd/amdkfd/kfd_device_queue_manager_vi.c | 2 +-
> drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c | 21 +++---
> drivers/gpu/drm/amd/amdkfd/kfd_events.c | 22 +++---
> drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c | 16 ++---
> drivers/gpu/drm/amd/amdkfd/kfd_module.c | 4 +-
> drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c | 10 ---
> drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c | 8 +--
> drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c | 34 ++++-----
> drivers/gpu/drm/amd/amdkfd/kfd_process.c | 4 +-
> .../gpu/drm/amd/amdkfd/kfd_process_queue_manager.c | 27 +++-----
> drivers/gpu/drm/amd/amdkfd/kfd_topology.c | 6 +-
> 17 files changed, 158 insertions(+), 236 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> index 98f4dbf..6244958 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> @@ -142,12 +142,12 @@ static int set_queue_properties_from_user(struct queue_properties *q_properties,
> struct kfd_ioctl_create_queue_args *args)
> {
> if (args->queue_percentage > KFD_MAX_QUEUE_PERCENTAGE) {
> - pr_err("kfd: queue percentage must be between 0 to KFD_MAX_QUEUE_PERCENTAGE\n");
> + pr_err("Queue percentage must be between 0 to KFD_MAX_QUEUE_PERCENTAGE\n");
> return -EINVAL;
> }
>
> if (args->queue_priority > KFD_MAX_QUEUE_PRIORITY) {
> - pr_err("kfd: queue priority must be between 0 to KFD_MAX_QUEUE_PRIORITY\n");
> + pr_err("Queue priority must be between 0 to KFD_MAX_QUEUE_PRIORITY\n");
> return -EINVAL;
> }
>
> @@ -155,26 +155,26 @@ static int set_queue_properties_from_user(struct queue_properties *q_properties,
> (!access_ok(VERIFY_WRITE,
> (const void __user *) args->ring_base_address,
> sizeof(uint64_t)))) {
> - pr_err("kfd: can't access ring base address\n");
> + pr_err("Can't access ring base address\n");
> return -EFAULT;
> }
>
> if (!is_power_of_2(args->ring_size) && (args->ring_size != 0)) {
> - pr_err("kfd: ring size must be a power of 2 or 0\n");
> + pr_err("Ring size must be a power of 2 or 0\n");
> return -EINVAL;
> }
>
> if (!access_ok(VERIFY_WRITE,
> (const void __user *) args->read_pointer_address,
> sizeof(uint32_t))) {
> - pr_err("kfd: can't access read pointer\n");
> + pr_err("Can't access read pointer\n");
> return -EFAULT;
> }
>
> if (!access_ok(VERIFY_WRITE,
> (const void __user *) args->write_pointer_address,
> sizeof(uint32_t))) {
> - pr_err("kfd: can't access write pointer\n");
> + pr_err("Can't access write pointer\n");
> return -EFAULT;
> }
>
> @@ -182,7 +182,7 @@ static int set_queue_properties_from_user(struct queue_properties *q_properties,
> !access_ok(VERIFY_WRITE,
> (const void __user *) args->eop_buffer_address,
> sizeof(uint32_t))) {
> - pr_debug("kfd: can't access eop buffer");
> + pr_debug("Can't access eop buffer");
> return -EFAULT;
> }
>
> @@ -190,7 +190,7 @@ static int set_queue_properties_from_user(struct queue_properties *q_properties,
> !access_ok(VERIFY_WRITE,
> (const void __user *) args->ctx_save_restore_address,
> sizeof(uint32_t))) {
> - pr_debug("kfd: can't access ctx save restore buffer");
> + pr_debug("Can't access ctx save restore buffer");
> return -EFAULT;
> }
>
> @@ -219,27 +219,27 @@ static int set_queue_properties_from_user(struct queue_properties *q_properties,
> else
> q_properties->format = KFD_QUEUE_FORMAT_PM4;
>
> - pr_debug("Queue Percentage (%d, %d)\n",
> + pr_debug("Queue Percentage: %d, %d\n",
> q_properties->queue_percent, args->queue_percentage);
>
> - pr_debug("Queue Priority (%d, %d)\n",
> + pr_debug("Queue Priority: %d, %d\n",
> q_properties->priority, args->queue_priority);
>
> - pr_debug("Queue Address (0x%llX, 0x%llX)\n",
> + pr_debug("Queue Address: 0x%llX, 0x%llX\n",
> q_properties->queue_address, args->ring_base_address);
>
> - pr_debug("Queue Size (0x%llX, %u)\n",
> + pr_debug("Queue Size: 0x%llX, %u\n",
> q_properties->queue_size, args->ring_size);
>
> - pr_debug("Queue r/w Pointers (0x%llX, 0x%llX)\n",
> - (uint64_t) q_properties->read_ptr,
> - (uint64_t) q_properties->write_ptr);
> + pr_debug("Queue r/w Pointers: %p, %p\n",
> + q_properties->read_ptr,
> + q_properties->write_ptr);
>
> - pr_debug("Queue Format (%d)\n", q_properties->format);
> + pr_debug("Queue Format: %d\n", q_properties->format);
>
> - pr_debug("Queue EOP (0x%llX)\n", q_properties->eop_ring_buffer_address);
> + pr_debug("Queue EOP: 0x%llX\n", q_properties->eop_ring_buffer_address);
>
> - pr_debug("Queue CTX save arex (0x%llX)\n",
> + pr_debug("Queue CTX save area: 0x%llX\n",
> q_properties->ctx_save_restore_area_address);
>
> return 0;
> @@ -257,16 +257,16 @@ static int kfd_ioctl_create_queue(struct file *filep, struct kfd_process *p,
>
> memset(&q_properties, 0, sizeof(struct queue_properties));
>
> - pr_debug("kfd: creating queue ioctl\n");
> + pr_debug("Creating queue ioctl\n");
>
> err = set_queue_properties_from_user(&q_properties, args);
> if (err)
> return err;
>
> - pr_debug("kfd: looking for gpu id 0x%x\n", args->gpu_id);
> + pr_debug("Looking for gpu id 0x%x\n", args->gpu_id);
> dev = kfd_device_by_id(args->gpu_id);
> if (dev == NULL) {
> - pr_debug("kfd: gpu id 0x%x was not found\n", args->gpu_id);
> + pr_debug("Could not find gpu id 0x%x\n", args->gpu_id);
> return -EINVAL;
> }
>
> @@ -278,7 +278,7 @@ static int kfd_ioctl_create_queue(struct file *filep, struct kfd_process *p,
> goto err_bind_process;
> }
>
> - pr_debug("kfd: creating queue for PASID %d on GPU 0x%x\n",
> + pr_debug("Creating queue for PASID %d on gpu 0x%x\n",
> p->pasid,
> dev->id);
>
> @@ -296,15 +296,15 @@ static int kfd_ioctl_create_queue(struct file *filep, struct kfd_process *p,
>
> mutex_unlock(&p->mutex);
>
> - pr_debug("kfd: queue id %d was created successfully\n", args->queue_id);
> + pr_debug("Queue id %d was created successfully\n", args->queue_id);
>
> - pr_debug("ring buffer address == 0x%016llX\n",
> + pr_debug("Ring buffer address == 0x%016llX\n",
> args->ring_base_address);
>
> - pr_debug("read ptr address == 0x%016llX\n",
> + pr_debug("Read ptr address == 0x%016llX\n",
> args->read_pointer_address);
>
> - pr_debug("write ptr address == 0x%016llX\n",
> + pr_debug("Write ptr address == 0x%016llX\n",
> args->write_pointer_address);
>
> return 0;
> @@ -321,7 +321,7 @@ static int kfd_ioctl_destroy_queue(struct file *filp, struct kfd_process *p,
> int retval;
> struct kfd_ioctl_destroy_queue_args *args = data;
>
> - pr_debug("kfd: destroying queue id %d for PASID %d\n",
> + pr_debug("Destroying queue id %d for pasid %d\n",
> args->queue_id,
> p->pasid);
>
> @@ -341,12 +341,12 @@ static int kfd_ioctl_update_queue(struct file *filp, struct kfd_process *p,
> struct queue_properties properties;
>
> if (args->queue_percentage > KFD_MAX_QUEUE_PERCENTAGE) {
> - pr_err("kfd: queue percentage must be between 0 to KFD_MAX_QUEUE_PERCENTAGE\n");
> + pr_err("Queue percentage must be between 0 to KFD_MAX_QUEUE_PERCENTAGE\n");
> return -EINVAL;
> }
>
> if (args->queue_priority > KFD_MAX_QUEUE_PRIORITY) {
> - pr_err("kfd: queue priority must be between 0 to KFD_MAX_QUEUE_PRIORITY\n");
> + pr_err("Queue priority must be between 0 to KFD_MAX_QUEUE_PRIORITY\n");
> return -EINVAL;
> }
>
> @@ -354,12 +354,12 @@ static int kfd_ioctl_update_queue(struct file *filp, struct kfd_process *p,
> (!access_ok(VERIFY_WRITE,
> (const void __user *) args->ring_base_address,
> sizeof(uint64_t)))) {
> - pr_err("kfd: can't access ring base address\n");
> + pr_err("Can't access ring base address\n");
> return -EFAULT;
> }
>
> if (!is_power_of_2(args->ring_size) && (args->ring_size != 0)) {
> - pr_err("kfd: ring size must be a power of 2 or 0\n");
> + pr_err("Ring size must be a power of 2 or 0\n");
> return -EINVAL;
> }
>
> @@ -368,7 +368,7 @@ static int kfd_ioctl_update_queue(struct file *filp, struct kfd_process *p,
> properties.queue_percent = args->queue_percentage;
> properties.priority = args->queue_priority;
>
> - pr_debug("kfd: updating queue id %d for PASID %d\n",
> + pr_debug("Updating queue id %d for pasid %d\n",
> args->queue_id, p->pasid);
>
> mutex_lock(&p->mutex);
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c b/drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c
> index a7548a5..bf8ee19 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c
> @@ -78,7 +78,7 @@ static int dbgdev_diq_submit_ib(struct kfd_dbgdev *dbgdev,
> pq_packets_size_in_bytes / sizeof(uint32_t),
> &ib_packet_buff);
> if (status != 0) {
> - pr_err("amdkfd: acquire_packet_buffer failed\n");
> + pr_err("acquire_packet_buffer failed\n");
> return status;
> }
>
> @@ -116,7 +116,7 @@ static int dbgdev_diq_submit_ib(struct kfd_dbgdev *dbgdev,
> &mem_obj);
>
> if (status != 0) {
> - pr_err("amdkfd: Failed to allocate GART memory\n");
> + pr_err("Failed to allocate GART memory\n");
> kq->ops.rollback_packet(kq);
> return status;
> }
> @@ -194,7 +194,7 @@ static int dbgdev_register_diq(struct kfd_dbgdev *dbgdev)
> &qid);
>
> if (status) {
> - pr_err("amdkfd: Failed to create DIQ\n");
> + pr_err("Failed to create DIQ\n");
> return status;
> }
>
> @@ -203,7 +203,7 @@ static int dbgdev_register_diq(struct kfd_dbgdev *dbgdev)
> kq = pqm_get_kernel_queue(dbgdev->pqm, qid);
>
> if (kq == NULL) {
> - pr_err("amdkfd: Error getting DIQ\n");
> + pr_err("Error getting DIQ\n");
> pqm_destroy_queue(dbgdev->pqm, qid);
> return -EFAULT;
> }
> @@ -279,7 +279,7 @@ static void dbgdev_address_watch_set_registers(
> }
>
> static int dbgdev_address_watch_nodiq(struct kfd_dbgdev *dbgdev,
> - struct dbg_address_watch_info *adw_info)
> + struct dbg_address_watch_info *adw_info)
> {
> union TCP_WATCH_ADDR_H_BITS addrHi;
> union TCP_WATCH_ADDR_L_BITS addrLo;
> @@ -293,7 +293,7 @@ static int dbgdev_address_watch_nodiq(struct kfd_dbgdev *dbgdev,
> pdd = kfd_get_process_device_data(dbgdev->dev,
> adw_info->process);
> if (!pdd) {
> - pr_err("amdkfd: Failed to get pdd for wave control no DIQ\n");
> + pr_err("Failed to get pdd for wave control no DIQ\n");
> return -EFAULT;
> }
>
> @@ -303,13 +303,13 @@ static int dbgdev_address_watch_nodiq(struct kfd_dbgdev *dbgdev,
>
> if ((adw_info->num_watch_points > MAX_WATCH_ADDRESSES) ||
> (adw_info->num_watch_points == 0)) {
> - pr_err("amdkfd: num_watch_points is invalid\n");
> + pr_err("num_watch_points is invalid\n");
> return -EINVAL;
> }
>
> if ((adw_info->watch_mode == NULL) ||
> (adw_info->watch_address == NULL)) {
> - pr_err("amdkfd: adw_info fields are not valid\n");
> + pr_err("adw_info fields are not valid\n");
> return -EINVAL;
> }
>
> @@ -348,7 +348,7 @@ static int dbgdev_address_watch_nodiq(struct kfd_dbgdev *dbgdev,
> }
>
> static int dbgdev_address_watch_diq(struct kfd_dbgdev *dbgdev,
> - struct dbg_address_watch_info *adw_info)
> + struct dbg_address_watch_info *adw_info)
> {
> struct pm4__set_config_reg *packets_vec;
> union TCP_WATCH_ADDR_H_BITS addrHi;
> @@ -371,20 +371,20 @@ static int dbgdev_address_watch_diq(struct kfd_dbgdev *dbgdev,
>
> if ((adw_info->num_watch_points > MAX_WATCH_ADDRESSES) ||
> (adw_info->num_watch_points == 0)) {
> - pr_err("amdkfd: num_watch_points is invalid\n");
> + pr_err("num_watch_points is invalid\n");
> return -EINVAL;
> }
>
> if ((NULL == adw_info->watch_mode) ||
> (NULL == adw_info->watch_address)) {
> - pr_err("amdkfd: adw_info fields are not valid\n");
> + pr_err("adw_info fields are not valid\n");
> return -EINVAL;
> }
>
> status = kfd_gtt_sa_allocate(dbgdev->dev, ib_size, &mem_obj);
>
> if (status != 0) {
> - pr_err("amdkfd: Failed to allocate GART memory\n");
> + pr_err("Failed to allocate GART memory\n");
> return status;
> }
>
> @@ -491,7 +491,7 @@ static int dbgdev_address_watch_diq(struct kfd_dbgdev *dbgdev,
> ib_size);
>
> if (status != 0) {
> - pr_err("amdkfd: Failed to submit IB to DIQ\n");
> + pr_err("Failed to submit IB to DIQ\n");
> break;
> }
> }
> @@ -619,7 +619,7 @@ static int dbgdev_wave_control_diq(struct kfd_dbgdev *dbgdev,
> status = dbgdev_wave_control_set_registers(wac_info, ®_sq_cmd,
> ®_gfx_index);
> if (status) {
> - pr_err("amdkfd: Failed to set wave control registers\n");
> + pr_err("Failed to set wave control registers\n");
> return status;
> }
>
> @@ -659,7 +659,7 @@ static int dbgdev_wave_control_diq(struct kfd_dbgdev *dbgdev,
> status = kfd_gtt_sa_allocate(dbgdev->dev, ib_size, &mem_obj);
>
> if (status != 0) {
> - pr_err("amdkfd: Failed to allocate GART memory\n");
> + pr_err("Failed to allocate GART memory\n");
> return status;
> }
>
> @@ -712,7 +712,7 @@ static int dbgdev_wave_control_diq(struct kfd_dbgdev *dbgdev,
> ib_size);
>
> if (status != 0)
> - pr_err("amdkfd: Failed to submit IB to DIQ\n");
> + pr_err("Failed to submit IB to DIQ\n");
>
> kfd_gtt_sa_free(dbgdev->dev, mem_obj);
>
> @@ -735,13 +735,13 @@ static int dbgdev_wave_control_nodiq(struct kfd_dbgdev *dbgdev,
> pdd = kfd_get_process_device_data(dbgdev->dev, wac_info->process);
>
> if (!pdd) {
> - pr_err("amdkfd: Failed to get pdd for wave control no DIQ\n");
> + pr_err("Failed to get pdd for wave control no DIQ\n");
> return -EFAULT;
> }
> status = dbgdev_wave_control_set_registers(wac_info, ®_sq_cmd,
> ®_gfx_index);
> if (status) {
> - pr_err("amdkfd: Failed to set wave control registers\n");
> + pr_err("Failed to set wave control registers\n");
> return status;
> }
>
> @@ -826,7 +826,7 @@ int dbgdev_wave_reset_wavefronts(struct kfd_dev *dev, struct kfd_process *p)
> }
>
> if (vmid > last_vmid_to_scan) {
> - pr_err("amdkfd: didn't found vmid for pasid (%d)\n", p->pasid);
> + pr_err("Didn't find vmid for pasid %d\n", p->pasid);
> return -EFAULT;
> }
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.c b/drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.c
> index 56d6763..7225789 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.c
> @@ -71,7 +71,7 @@ bool kfd_dbgmgr_create(struct kfd_dbgmgr **ppmgr, struct kfd_dev *pdev)
>
> new_buff = kfd_alloc_struct(new_buff);
> if (!new_buff) {
> - pr_err("amdkfd: Failed to allocate dbgmgr instance\n");
> + pr_err("Failed to allocate dbgmgr instance\n");
> return false;
> }
>
> @@ -79,7 +79,7 @@ bool kfd_dbgmgr_create(struct kfd_dbgmgr **ppmgr, struct kfd_dev *pdev)
> new_buff->dev = pdev;
> new_buff->dbgdev = kfd_alloc_struct(new_buff->dbgdev);
> if (!new_buff->dbgdev) {
> - pr_err("amdkfd: Failed to allocate dbgdev instance\n");
> + pr_err("Failed to allocate dbgdev instance\n");
> kfree(new_buff);
> return false;
> }
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device.c b/drivers/gpu/drm/amd/amdkfd/kfd_device.c
> index 1f50325..87df8bf 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device.c
> @@ -152,7 +152,7 @@ static bool device_iommu_pasid_init(struct kfd_dev *kfd)
> }
>
> if ((iommu_info.flags & required_iommu_flags) != required_iommu_flags) {
> - dev_err(kfd_device, "error required iommu flags ats(%i), pri(%i), pasid(%i)\n",
> + dev_err(kfd_device, "error required iommu flags ats %i, pri %i, pasid %i\n",
> (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_ATS_SUP) != 0,
> (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PRI_SUP) != 0,
> (iommu_info.flags & AMD_IOMMU_DEVICE_FLAG_PASID_SUP)
> @@ -248,42 +248,33 @@ bool kgd2kfd_device_init(struct kfd_dev *kfd,
> if (kfd->kfd2kgd->init_gtt_mem_allocation(
> kfd->kgd, size, &kfd->gtt_mem,
> &kfd->gtt_start_gpu_addr, &kfd->gtt_start_cpu_ptr)){
> - dev_err(kfd_device,
> - "Could not allocate %d bytes for device (%x:%x)\n",
> - size, kfd->pdev->vendor, kfd->pdev->device);
> + dev_err(kfd_device, "Could not allocate %d bytes\n", size);
> goto out;
> }
>
> - dev_info(kfd_device,
> - "Allocated %d bytes on gart for device(%x:%x)\n",
> - size, kfd->pdev->vendor, kfd->pdev->device);
> + dev_info(kfd_device, "Allocated %d bytes on gart\n", size);
>
> /* Initialize GTT sa with 512 byte chunk size */
> if (kfd_gtt_sa_init(kfd, size, 512) != 0) {
> - dev_err(kfd_device,
> - "Error initializing gtt sub-allocator\n");
> + dev_err(kfd_device, "Error initializing gtt sub-allocator\n");
> goto kfd_gtt_sa_init_error;
> }
>
> kfd_doorbell_init(kfd);
>
> if (kfd_topology_add_device(kfd) != 0) {
> - dev_err(kfd_device,
> - "Error adding device (%x:%x) to topology\n",
> - kfd->pdev->vendor, kfd->pdev->device);
> + dev_err(kfd_device, "Error adding device to topology\n");
> goto kfd_topology_add_device_error;
> }
>
> if (kfd_interrupt_init(kfd)) {
> - dev_err(kfd_device,
> - "Error initializing interrupts for device (%x:%x)\n",
> - kfd->pdev->vendor, kfd->pdev->device);
> + dev_err(kfd_device, "Error initializing interrupts\n");
> goto kfd_interrupt_error;
> }
>
> if (!device_iommu_pasid_init(kfd)) {
> dev_err(kfd_device,
> - "Error initializing iommuv2 for device (%x:%x)\n",
> + "Error initializing iommuv2 for device %x:%x\n",
> kfd->pdev->vendor, kfd->pdev->device);
> goto device_iommu_pasid_error;
> }
> @@ -293,15 +284,13 @@ bool kgd2kfd_device_init(struct kfd_dev *kfd,
>
> kfd->dqm = device_queue_manager_init(kfd);
> if (!kfd->dqm) {
> - dev_err(kfd_device,
> - "Error initializing queue manager for device (%x:%x)\n",
> - kfd->pdev->vendor, kfd->pdev->device);
> + dev_err(kfd_device, "Error initializing queue manager\n");
> goto device_queue_manager_error;
> }
>
> if (kfd->dqm->ops.start(kfd->dqm) != 0) {
> dev_err(kfd_device,
> - "Error starting queuen manager for device (%x:%x)\n",
> + "Error starting queue manager for device %x:%x\n",
> kfd->pdev->vendor, kfd->pdev->device);
> goto dqm_start_error;
> }
> @@ -309,10 +298,10 @@ bool kgd2kfd_device_init(struct kfd_dev *kfd,
> kfd->dbgmgr = NULL;
>
> kfd->init_complete = true;
> - dev_info(kfd_device, "added device (%x:%x)\n", kfd->pdev->vendor,
> + dev_info(kfd_device, "added device %x:%x\n", kfd->pdev->vendor,
> kfd->pdev->device);
>
> - pr_debug("kfd: Starting kfd with the following scheduling policy %d\n",
> + pr_debug("Starting kfd with the following scheduling policy %d\n",
> sched_policy);
>
> goto out;
> @@ -330,7 +319,7 @@ bool kgd2kfd_device_init(struct kfd_dev *kfd,
> kfd_gtt_sa_init_error:
> kfd->kfd2kgd->free_gtt_mem(kfd->kgd, kfd->gtt_mem);
> dev_err(kfd_device,
> - "device (%x:%x) NOT added due to errors\n",
> + "device %x:%x NOT added due to errors\n",
> kfd->pdev->vendor, kfd->pdev->device);
> out:
> return kfd->init_complete;
> @@ -422,7 +411,7 @@ static int kfd_gtt_sa_init(struct kfd_dev *kfd, unsigned int buf_size,
> if (!kfd->gtt_sa_bitmap)
> return -ENOMEM;
>
> - pr_debug("kfd: gtt_sa_num_of_chunks = %d, gtt_sa_bitmap = %p\n",
> + pr_debug("gtt_sa_num_of_chunks = %d, gtt_sa_bitmap = %p\n",
> kfd->gtt_sa_num_of_chunks, kfd->gtt_sa_bitmap);
>
> mutex_init(&kfd->gtt_sa_lock);
> @@ -468,7 +457,7 @@ int kfd_gtt_sa_allocate(struct kfd_dev *kfd, unsigned int size,
> if ((*mem_obj) == NULL)
> return -ENOMEM;
>
> - pr_debug("kfd: allocated mem_obj = %p for size = %d\n", *mem_obj, size);
> + pr_debug("Allocated mem_obj = %p for size = %d\n", *mem_obj, size);
>
> start_search = 0;
>
> @@ -480,7 +469,7 @@ int kfd_gtt_sa_allocate(struct kfd_dev *kfd, unsigned int size,
> kfd->gtt_sa_num_of_chunks,
> start_search);
>
> - pr_debug("kfd: found = %d\n", found);
> + pr_debug("Found = %d\n", found);
>
> /* If there wasn't any free chunk, bail out */
> if (found == kfd->gtt_sa_num_of_chunks)
> @@ -498,12 +487,12 @@ int kfd_gtt_sa_allocate(struct kfd_dev *kfd, unsigned int size,
> found,
> kfd->gtt_sa_chunk_size);
>
> - pr_debug("kfd: gpu_addr = %p, cpu_addr = %p\n",
> + pr_debug("gpu_addr = %p, cpu_addr = %p\n",
> (uint64_t *) (*mem_obj)->gpu_addr, (*mem_obj)->cpu_ptr);
>
> /* If we need only one chunk, mark it as allocated and get out */
> if (size <= kfd->gtt_sa_chunk_size) {
> - pr_debug("kfd: single bit\n");
> + pr_debug("Single bit\n");
> set_bit(found, kfd->gtt_sa_bitmap);
> goto kfd_gtt_out;
> }
> @@ -538,7 +527,7 @@ int kfd_gtt_sa_allocate(struct kfd_dev *kfd, unsigned int size,
>
> } while (cur_size > 0);
>
> - pr_debug("kfd: range_start = %d, range_end = %d\n",
> + pr_debug("range_start = %d, range_end = %d\n",
> (*mem_obj)->range_start, (*mem_obj)->range_end);
>
> /* Mark the chunks as allocated */
> @@ -552,7 +541,7 @@ int kfd_gtt_sa_allocate(struct kfd_dev *kfd, unsigned int size,
> return 0;
>
> kfd_gtt_no_free_chunk:
> - pr_debug("kfd: allocation failed with mem_obj = %p\n", mem_obj);
> + pr_debug("Allocation failed with mem_obj = %p\n", mem_obj);
> mutex_unlock(&kfd->gtt_sa_lock);
> kfree(mem_obj);
> return -ENOMEM;
> @@ -568,7 +557,7 @@ int kfd_gtt_sa_free(struct kfd_dev *kfd, struct kfd_mem_obj *mem_obj)
> if (!mem_obj)
> return 0;
>
> - pr_debug("kfd: free mem_obj = %p, range_start = %d, range_end = %d\n",
> + pr_debug("Free mem_obj = %p, range_start = %d, range_end = %d\n",
> mem_obj, mem_obj->range_start, mem_obj->range_end);
>
> mutex_lock(&kfd->gtt_sa_lock);
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> index 3b850da..8b147e4 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> @@ -121,7 +121,7 @@ static int allocate_vmid(struct device_queue_manager *dqm,
>
> /* Kaveri kfd vmid's starts from vmid 8 */
> allocated_vmid = bit + KFD_VMID_START_OFFSET;
> - pr_debug("kfd: vmid allocation %d\n", allocated_vmid);
> + pr_debug("vmid allocation %d\n", allocated_vmid);
> qpd->vmid = allocated_vmid;
> q->properties.vmid = allocated_vmid;
>
> @@ -154,13 +154,12 @@ static int create_queue_nocpsch(struct device_queue_manager *dqm,
>
> BUG_ON(!dqm || !q || !qpd || !allocated_vmid);
>
> - pr_debug("kfd: In func %s\n", __func__);
> print_queue(q);
>
> mutex_lock(&dqm->lock);
>
> if (dqm->total_queue_count >= max_num_of_queues_per_device) {
> - pr_warn("amdkfd: Can't create new usermode queue because %d queues were already created\n",
> + pr_warn("Can't create new usermode queue because %d queues were already created\n",
> dqm->total_queue_count);
> mutex_unlock(&dqm->lock);
> return -EPERM;
> @@ -240,8 +239,7 @@ static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q)
> if (!set)
> return -EBUSY;
>
> - pr_debug("kfd: DQM %s hqd slot - pipe (%d) queue(%d)\n",
> - __func__, q->pipe, q->queue);
> + pr_debug("hqd slot - pipe %d, queue %d\n", q->pipe, q->queue);
> /* horizontal hqd allocation */
> dqm->next_pipe_to_allocate = (pipe + 1) % get_pipes_per_mec(dqm);
>
> @@ -278,9 +276,8 @@ static int create_compute_queue_nocpsch(struct device_queue_manager *dqm,
> return retval;
> }
>
> - pr_debug("kfd: loading mqd to hqd on pipe (%d) queue (%d)\n",
> - q->pipe,
> - q->queue);
> + pr_debug("Loading mqd to hqd on pipe %d, queue %d\n",
> + q->pipe, q->queue);
>
> retval = mqd->load_mqd(mqd, q->mqd, q->pipe,
> q->queue, (uint32_t __user *) q->properties.write_ptr);
> @@ -304,8 +301,6 @@ static int destroy_queue_nocpsch(struct device_queue_manager *dqm,
>
> retval = 0;
>
> - pr_debug("kfd: In Func %s\n", __func__);
> -
> mutex_lock(&dqm->lock);
>
> if (q->properties.type == KFD_QUEUE_TYPE_COMPUTE) {
> @@ -324,7 +319,7 @@ static int destroy_queue_nocpsch(struct device_queue_manager *dqm,
> dqm->sdma_queue_count--;
> deallocate_sdma_queue(dqm, q->sdma_id);
> } else {
> - pr_debug("q->properties.type is invalid (%d)\n",
> + pr_debug("q->properties.type %d is invalid\n",
> q->properties.type);
> retval = -EINVAL;
> goto out;
> @@ -403,13 +398,13 @@ static struct mqd_manager *get_mqd_manager_nocpsch(
>
> BUG_ON(!dqm || type >= KFD_MQD_TYPE_MAX);
>
> - pr_debug("kfd: In func %s mqd type %d\n", __func__, type);
> + pr_debug("mqd type %d\n", type);
>
> mqd = dqm->mqds[type];
> if (!mqd) {
> mqd = mqd_manager_init(type, dqm->dev);
> if (mqd == NULL)
> - pr_err("kfd: mqd manager is NULL");
> + pr_err("mqd manager is NULL");
> dqm->mqds[type] = mqd;
> }
>
> @@ -424,8 +419,6 @@ static int register_process_nocpsch(struct device_queue_manager *dqm,
>
> BUG_ON(!dqm || !qpd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> n = kzalloc(sizeof(struct device_process_node), GFP_KERNEL);
> if (!n)
> return -ENOMEM;
> @@ -452,8 +445,6 @@ static int unregister_process_nocpsch(struct device_queue_manager *dqm,
>
> BUG_ON(!dqm || !qpd);
>
> - pr_debug("In func %s\n", __func__);
> -
> pr_debug("qpd->queues_list is %s\n",
> list_empty(&qpd->queues_list) ? "empty" : "not empty");
>
> @@ -501,25 +492,13 @@ static void init_interrupts(struct device_queue_manager *dqm)
> dqm->dev->kfd2kgd->init_interrupts(dqm->dev->kgd, i);
> }
>
> -static int init_scheduler(struct device_queue_manager *dqm)
> -{
> - int retval = 0;
> -
> - BUG_ON(!dqm);
> -
> - pr_debug("kfd: In %s\n", __func__);
> -
> - return retval;
> -}
> -
> static int initialize_nocpsch(struct device_queue_manager *dqm)
> {
> int pipe, queue;
>
> BUG_ON(!dqm);
>
> - pr_debug("kfd: In func %s num of pipes: %d\n",
> - __func__, get_pipes_per_mec(dqm));
> + pr_debug("num of pipes: %d\n", get_pipes_per_mec(dqm));
>
> mutex_init(&dqm->lock);
> INIT_LIST_HEAD(&dqm->queues);
> @@ -544,7 +523,6 @@ static int initialize_nocpsch(struct device_queue_manager *dqm)
> dqm->vmid_bitmap = (1 << VMID_PER_DEVICE) - 1;
> dqm->sdma_bitmap = (1 << CIK_SDMA_QUEUES) - 1;
>
> - init_scheduler(dqm);
> return 0;
> }
>
> @@ -617,9 +595,9 @@ static int create_sdma_queue_nocpsch(struct device_queue_manager *dqm,
> q->properties.sdma_queue_id = q->sdma_id % CIK_SDMA_QUEUES_PER_ENGINE;
> q->properties.sdma_engine_id = q->sdma_id / CIK_SDMA_ENGINE_NUM;
>
> - pr_debug("kfd: sdma id is: %d\n", q->sdma_id);
> - pr_debug(" sdma queue id: %d\n", q->properties.sdma_queue_id);
> - pr_debug(" sdma engine id: %d\n", q->properties.sdma_engine_id);
> + pr_debug("SDMA id is: %d\n", q->sdma_id);
> + pr_debug("SDMA queue id: %d\n", q->properties.sdma_queue_id);
> + pr_debug("SDMA engine id: %d\n", q->properties.sdma_engine_id);
>
> dqm->ops_asic_specific.init_sdma_vm(dqm, q, qpd);
> retval = mqd->init_mqd(mqd, &q->mqd, &q->mqd_mem_obj,
> @@ -651,8 +629,6 @@ static int set_sched_resources(struct device_queue_manager *dqm)
>
> BUG_ON(!dqm);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> res.vmid_mask = (1 << VMID_PER_DEVICE) - 1;
> res.vmid_mask <<= KFD_VMID_START_OFFSET;
>
> @@ -682,9 +658,9 @@ static int set_sched_resources(struct device_queue_manager *dqm)
> res.gws_mask = res.oac_mask = res.gds_heap_base =
> res.gds_heap_size = 0;
>
> - pr_debug("kfd: scheduling resources:\n"
> - " vmid mask: 0x%8X\n"
> - " queue mask: 0x%8llX\n",
> + pr_debug("Scheduling resources:\n"
> + "vmid mask: 0x%8X\n"
> + "queue mask: 0x%8llX\n",
> res.vmid_mask, res.queue_mask);
>
> return pm_send_set_resources(&dqm->packets, &res);
> @@ -696,8 +672,7 @@ static int initialize_cpsch(struct device_queue_manager *dqm)
>
> BUG_ON(!dqm);
>
> - pr_debug("kfd: In func %s num of pipes: %d\n",
> - __func__, get_pipes_per_mec(dqm));
> + pr_debug("num of pipes: %d\n", get_pipes_per_mec(dqm));
>
> mutex_init(&dqm->lock);
> INIT_LIST_HEAD(&dqm->queues);
> @@ -732,7 +707,7 @@ static int start_cpsch(struct device_queue_manager *dqm)
> if (retval != 0)
> goto fail_set_sched_resources;
>
> - pr_debug("kfd: allocating fence memory\n");
> + pr_debug("Allocating fence memory\n");
>
> /* allocate fence memory on the gart */
> retval = kfd_gtt_sa_allocate(dqm->dev, sizeof(*dqm->fence_addr),
> @@ -786,11 +761,9 @@ static int create_kernel_queue_cpsch(struct device_queue_manager *dqm,
> {
> BUG_ON(!dqm || !kq || !qpd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> mutex_lock(&dqm->lock);
> if (dqm->total_queue_count >= max_num_of_queues_per_device) {
> - pr_warn("amdkfd: Can't create new kernel queue because %d queues were already created\n",
> + pr_warn("Can't create new kernel queue because %d queues were already created\n",
> dqm->total_queue_count);
> mutex_unlock(&dqm->lock);
> return -EPERM;
> @@ -819,8 +792,6 @@ static void destroy_kernel_queue_cpsch(struct device_queue_manager *dqm,
> {
> BUG_ON(!dqm || !kq);
>
> - pr_debug("kfd: In %s\n", __func__);
> -
> mutex_lock(&dqm->lock);
> /* here we actually preempt the DIQ */
> destroy_queues_cpsch(dqm, true, false);
> @@ -862,7 +833,7 @@ static int create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q,
> mutex_lock(&dqm->lock);
>
> if (dqm->total_queue_count >= max_num_of_queues_per_device) {
> - pr_warn("amdkfd: Can't create new usermode queue because %d queues were already created\n",
> + pr_warn("Can't create new usermode queue because %d queues were already created\n",
> dqm->total_queue_count);
> retval = -EPERM;
> goto out;
> @@ -916,7 +887,7 @@ int amdkfd_fence_wait_timeout(unsigned int *fence_addr,
>
> while (*fence_addr != fence_value) {
> if (time_after(jiffies, timeout)) {
> - pr_err("kfd: qcm fence wait loop timeout expired\n");
> + pr_err("qcm fence wait loop timeout expired\n");
> return -ETIME;
> }
> schedule();
> @@ -949,7 +920,7 @@ static int destroy_queues_cpsch(struct device_queue_manager *dqm,
> if (!dqm->active_runlist)
> goto out;
>
> - pr_debug("kfd: Before destroying queues, sdma queue count is : %u\n",
> + pr_debug("Before destroying queues, sdma queue count is : %u\n",
> dqm->sdma_queue_count);
>
> if (dqm->sdma_queue_count > 0) {
> @@ -998,7 +969,7 @@ static int execute_queues_cpsch(struct device_queue_manager *dqm, bool lock)
>
> retval = destroy_queues_cpsch(dqm, false, false);
> if (retval != 0) {
> - pr_err("kfd: the cp might be in an unrecoverable state due to an unsuccessful queues preemption");
> + pr_err("The cp might be in an unrecoverable state due to an unsuccessful queues preemption");
> goto out;
> }
>
> @@ -1014,7 +985,7 @@ static int execute_queues_cpsch(struct device_queue_manager *dqm, bool lock)
>
> retval = pm_send_runlist(&dqm->packets, &dqm->queues);
> if (retval != 0) {
> - pr_err("kfd: failed to execute runlist");
> + pr_err("failed to execute runlist");
> goto out;
> }
> dqm->active_runlist = true;
> @@ -1106,8 +1077,6 @@ static bool set_cache_memory_policy(struct device_queue_manager *dqm,
> {
> bool retval;
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> mutex_lock(&dqm->lock);
>
> if (alternate_aperture_size == 0) {
> @@ -1152,7 +1121,7 @@ static bool set_cache_memory_policy(struct device_queue_manager *dqm,
> if ((sched_policy == KFD_SCHED_POLICY_NO_HWS) && (qpd->vmid != 0))
> program_sh_mem_settings(dqm, qpd);
>
> - pr_debug("kfd: sh_mem_config: 0x%x, ape1_base: 0x%x, ape1_limit: 0x%x\n",
> + pr_debug("sh_mem_config: 0x%x, ape1_base: 0x%x, ape1_limit: 0x%x\n",
> qpd->sh_mem_config, qpd->sh_mem_ape1_base,
> qpd->sh_mem_ape1_limit);
>
> @@ -1170,7 +1139,7 @@ struct device_queue_manager *device_queue_manager_init(struct kfd_dev *dev)
>
> BUG_ON(!dev);
>
> - pr_debug("kfd: loading device queue manager\n");
> + pr_debug("Loading device queue manager\n");
>
> dqm = kzalloc(sizeof(struct device_queue_manager), GFP_KERNEL);
> if (!dqm)
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
> index 48dc056..a263e2a 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
> @@ -127,7 +127,7 @@ static int register_process_cik(struct device_queue_manager *dqm,
> qpd->sh_mem_bases = compute_sh_mem_bases_64bit(temp);
> }
>
> - pr_debug("kfd: is32bit process: %d sh_mem_bases nybble: 0x%X and register 0x%X\n",
> + pr_debug("is32bit process: %d sh_mem_bases nybble: 0x%X and register 0x%X\n",
> qpd->pqm->process->is_32bit_user_mode, temp, qpd->sh_mem_bases);
>
> return 0;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
> index 7e9cae9..8c45c86 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
> @@ -139,7 +139,7 @@ static int register_process_vi(struct device_queue_manager *dqm,
> SH_MEM_CONFIG__ADDRESS_MODE__SHIFT;
> }
>
> - pr_debug("kfd: is32bit process: %d sh_mem_bases nybble: 0x%X and register 0x%X\n",
> + pr_debug("is32bit process: %d sh_mem_bases nybble: 0x%X and register 0x%X\n",
> qpd->pqm->process->is_32bit_user_mode, temp, qpd->sh_mem_bases);
>
> return 0;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c b/drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c
> index 453c5d6..ca21538 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c
> @@ -97,23 +97,23 @@ void kfd_doorbell_init(struct kfd_dev *kfd)
>
> BUG_ON(!kfd->doorbell_kernel_ptr);
>
> - pr_debug("kfd: doorbell initialization:\n");
> - pr_debug("kfd: doorbell base == 0x%08lX\n",
> + pr_debug("Doorbell initialization:\n");
> + pr_debug("doorbell base == 0x%08lX\n",
> (uintptr_t)kfd->doorbell_base);
>
> - pr_debug("kfd: doorbell_id_offset == 0x%08lX\n",
> + pr_debug("doorbell_id_offset == 0x%08lX\n",
> kfd->doorbell_id_offset);
>
> - pr_debug("kfd: doorbell_process_limit == 0x%08lX\n",
> + pr_debug("doorbell_process_limit == 0x%08lX\n",
> doorbell_process_limit);
>
> - pr_debug("kfd: doorbell_kernel_offset == 0x%08lX\n",
> + pr_debug("doorbell_kernel_offset == 0x%08lX\n",
> (uintptr_t)kfd->doorbell_base);
>
> - pr_debug("kfd: doorbell aperture size == 0x%08lX\n",
> + pr_debug("doorbell aperture size == 0x%08lX\n",
> kfd->shared_resources.doorbell_aperture_size);
>
> - pr_debug("kfd: doorbell kernel address == 0x%08lX\n",
> + pr_debug("doorbell kernel address == 0x%08lX\n",
> (uintptr_t)kfd->doorbell_kernel_ptr);
> }
>
> @@ -142,12 +142,11 @@ int kfd_doorbell_mmap(struct kfd_process *process, struct vm_area_struct *vma)
>
> vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
>
> - pr_debug("kfd: mapping doorbell page in %s\n"
> + pr_debug("Mapping doorbell page\n"
> " target user address == 0x%08llX\n"
> " physical address == 0x%08llX\n"
> " vm_flags == 0x%04lX\n"
> " size == 0x%04lX\n",
> - __func__,
> (unsigned long long) vma->vm_start, address, vma->vm_flags,
> doorbell_process_allocation());
>
> @@ -185,7 +184,7 @@ u32 __iomem *kfd_get_kernel_doorbell(struct kfd_dev *kfd,
> *doorbell_off = KERNEL_DOORBELL_PASID * (doorbell_process_allocation() /
> sizeof(u32)) + inx;
>
> - pr_debug("kfd: get kernel queue doorbell\n"
> + pr_debug("Get kernel queue doorbell\n"
> " doorbell offset == 0x%08X\n"
> " kernel address == 0x%08lX\n",
> *doorbell_off, (uintptr_t)(kfd->doorbell_kernel_ptr + inx));
> @@ -210,7 +209,7 @@ inline void write_kernel_doorbell(u32 __iomem *db, u32 value)
> {
> if (db) {
> writel(value, db);
> - pr_debug("writing %d to doorbell address 0x%p\n", value, db);
> + pr_debug("Writing %d to doorbell address 0x%p\n", value, db);
> }
> }
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_events.c b/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> index d8b9b3c..abdaf95 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_events.c
> @@ -110,7 +110,7 @@ static bool allocate_free_slot(struct kfd_process *process,
> *out_page = page;
> *out_slot_index = slot;
>
> - pr_debug("allocated event signal slot in page %p, slot %d\n",
> + pr_debug("Allocated event signal slot in page %p, slot %d\n",
> page, slot);
>
> return true;
> @@ -155,9 +155,9 @@ static bool allocate_signal_page(struct file *devkfd, struct kfd_process *p)
> struct signal_page,
> event_pages)->page_index + 1;
>
> - pr_debug("allocated new event signal page at %p, for process %p\n",
> + pr_debug("Allocated new event signal page at %p, for process %p\n",
> page, p);
> - pr_debug("page index is %d\n", page->page_index);
> + pr_debug("Page index is %d\n", page->page_index);
>
> list_add(&page->event_pages, &p->signal_event_pages);
>
> @@ -292,13 +292,13 @@ static int create_signal_event(struct file *devkfd,
> struct kfd_event *ev)
> {
> if (p->signal_event_count == KFD_SIGNAL_EVENT_LIMIT) {
> - pr_warn("amdkfd: Signal event wasn't created because limit was reached\n");
> + pr_warn("Signal event wasn't created because limit was reached\n");
> return -ENOMEM;
> }
>
> if (!allocate_event_notification_slot(devkfd, p, &ev->signal_page,
> &ev->signal_slot_index)) {
> - pr_warn("amdkfd: Signal event wasn't created because out of kernel memory\n");
> + pr_warn("Signal event wasn't created because out of kernel memory\n");
> return -ENOMEM;
> }
>
> @@ -310,11 +310,7 @@ static int create_signal_event(struct file *devkfd,
> ev->event_id = make_signal_event_id(ev->signal_page,
> ev->signal_slot_index);
>
> - pr_debug("signal event number %zu created with id %d, address %p\n",
> - p->signal_event_count, ev->event_id,
> - ev->user_signal_address);
> -
> - pr_debug("signal event number %zu created with id %d, address %p\n",
> + pr_debug("Signal event number %zu created with id %d, address %p\n",
> p->signal_event_count, ev->event_id,
> ev->user_signal_address);
>
> @@ -817,7 +813,7 @@ int kfd_event_mmap(struct kfd_process *p, struct vm_area_struct *vma)
> /* check required size is logical */
> if (get_order(KFD_SIGNAL_EVENT_LIMIT * 8) !=
> get_order(vma->vm_end - vma->vm_start)) {
> - pr_err("amdkfd: event page mmap requested illegal size\n");
> + pr_err("Event page mmap requested illegal size\n");
> return -EINVAL;
> }
>
> @@ -826,7 +822,7 @@ int kfd_event_mmap(struct kfd_process *p, struct vm_area_struct *vma)
> page = lookup_signal_page_by_index(p, page_index);
> if (!page) {
> /* Probably KFD bug, but mmap is user-accessible. */
> - pr_debug("signal page could not be found for page_index %u\n",
> + pr_debug("Signal page could not be found for page_index %u\n",
> page_index);
> return -EINVAL;
> }
> @@ -837,7 +833,7 @@ int kfd_event_mmap(struct kfd_process *p, struct vm_area_struct *vma)
> vma->vm_flags |= VM_IO | VM_DONTCOPY | VM_DONTEXPAND | VM_NORESERVE
> | VM_DONTDUMP | VM_PFNMAP;
>
> - pr_debug("mapping signal page\n");
> + pr_debug("Mapping signal page\n");
> pr_debug(" start user address == 0x%08lx\n", vma->vm_start);
> pr_debug(" end user address == 0x%08lx\n", vma->vm_end);
> pr_debug(" pfn == 0x%016lX\n", pfn);
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c b/drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
> index d135cd0..f89d366 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
> @@ -44,8 +44,8 @@ static bool initialize(struct kernel_queue *kq, struct kfd_dev *dev,
> BUG_ON(!kq || !dev);
> BUG_ON(type != KFD_QUEUE_TYPE_DIQ && type != KFD_QUEUE_TYPE_HIQ);
>
> - pr_debug("amdkfd: In func %s initializing queue type %d size %d\n",
> - __func__, KFD_QUEUE_TYPE_HIQ, queue_size);
> + pr_debug("Initializing queue type %d size %d\n", KFD_QUEUE_TYPE_HIQ,
> + queue_size);
>
> memset(&prop, 0, sizeof(prop));
> memset(&nop, 0, sizeof(nop));
> @@ -73,13 +73,13 @@ static bool initialize(struct kernel_queue *kq, struct kfd_dev *dev,
> prop.doorbell_ptr = kfd_get_kernel_doorbell(dev, &prop.doorbell_off);
>
> if (prop.doorbell_ptr == NULL) {
> - pr_err("amdkfd: error init doorbell");
> + pr_err("Failed to initialize doorbell");
> goto err_get_kernel_doorbell;
> }
>
> retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq);
> if (retval != 0) {
> - pr_err("amdkfd: error init pq queues size (%d)\n", queue_size);
> + pr_err("Failed to init pq queues size %d\n", queue_size);
> goto err_pq_allocate_vidmem;
> }
>
> @@ -139,7 +139,7 @@ static bool initialize(struct kernel_queue *kq, struct kfd_dev *dev,
>
> /* assign HIQ to HQD */
> if (type == KFD_QUEUE_TYPE_HIQ) {
> - pr_debug("assigning hiq to hqd\n");
> + pr_debug("Assigning hiq to hqd\n");
> kq->queue->pipe = KFD_CIK_HIQ_PIPE;
> kq->queue->queue = KFD_CIK_HIQ_QUEUE;
> kq->mqd->load_mqd(kq->mqd, kq->queue->mqd, kq->queue->pipe,
> @@ -304,7 +304,7 @@ struct kernel_queue *kernel_queue_init(struct kfd_dev *dev,
> }
>
> if (!kq->ops.initialize(kq, dev, type, KFD_KERNEL_QUEUE_SIZE)) {
> - pr_err("amdkfd: failed to init kernel queue\n");
> + pr_err("Failed to init kernel queue\n");
> kfree(kq);
> return NULL;
> }
> @@ -327,7 +327,7 @@ static __attribute__((unused)) void test_kq(struct kfd_dev *dev)
>
> BUG_ON(!dev);
>
> - pr_err("amdkfd: starting kernel queue test\n");
> + pr_err("Starting kernel queue test\n");
>
> kq = kernel_queue_init(dev, KFD_QUEUE_TYPE_HIQ);
> BUG_ON(!kq);
> @@ -338,7 +338,7 @@ static __attribute__((unused)) void test_kq(struct kfd_dev *dev)
> buffer[i] = kq->nop_packet;
> kq->ops.submit_packet(kq);
>
> - pr_err("amdkfd: ending kernel queue test\n");
> + pr_err("Ending kernel queue test\n");
> }
>
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_module.c b/drivers/gpu/drm/amd/amdkfd/kfd_module.c
> index af5bfc1..819a442 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_module.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_module.c
> @@ -91,7 +91,7 @@ static int __init kfd_module_init(void)
> /* Verify module parameters */
> if ((sched_policy < KFD_SCHED_POLICY_HWS) ||
> (sched_policy > KFD_SCHED_POLICY_NO_HWS)) {
> - pr_err("kfd: sched_policy has invalid value\n");
> + pr_err("sched_policy has invalid value\n");
> return -1;
> }
>
> @@ -99,7 +99,7 @@ static int __init kfd_module_init(void)
> if ((max_num_of_queues_per_device < 1) ||
> (max_num_of_queues_per_device >
> KFD_MAX_NUM_OF_QUEUES_PER_DEVICE)) {
> - pr_err("kfd: max_num_of_queues_per_device must be between 1 to KFD_MAX_NUM_OF_QUEUES_PER_DEVICE\n");
> + pr_err("max_num_of_queues_per_device must be between 1 to KFD_MAX_NUM_OF_QUEUES_PER_DEVICE\n");
> return -1;
> }
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> index ac59229..27fd930 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> @@ -46,8 +46,6 @@ static int init_mqd(struct mqd_manager *mm, void **mqd,
>
> BUG_ON(!mm || !q || !mqd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> retval = kfd_gtt_sa_allocate(mm->dev, sizeof(struct cik_mqd),
> mqd_mem_obj);
>
> @@ -172,8 +170,6 @@ static int update_mqd(struct mqd_manager *mm, void *mqd,
>
> BUG_ON(!mm || !q || !mqd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> m = get_mqd(mqd);
> m->cp_hqd_pq_control = DEFAULT_RPTR_BLOCK_SIZE |
> DEFAULT_MIN_AVAIL_SIZE | PQ_ATC_EN;
> @@ -302,8 +298,6 @@ static int init_mqd_hiq(struct mqd_manager *mm, void **mqd,
>
> BUG_ON(!mm || !q || !mqd || !mqd_mem_obj);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> retval = kfd_gtt_sa_allocate(mm->dev, sizeof(struct cik_mqd),
> mqd_mem_obj);
>
> @@ -360,8 +354,6 @@ static int update_mqd_hiq(struct mqd_manager *mm, void *mqd,
>
> BUG_ON(!mm || !q || !mqd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> m = get_mqd(mqd);
> m->cp_hqd_pq_control = DEFAULT_RPTR_BLOCK_SIZE |
> DEFAULT_MIN_AVAIL_SIZE |
> @@ -414,8 +406,6 @@ struct mqd_manager *mqd_manager_init_cik(enum KFD_MQD_TYPE type,
> BUG_ON(!dev);
> BUG_ON(type >= KFD_MQD_TYPE_MAX);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> mqd = kzalloc(sizeof(struct mqd_manager), GFP_KERNEL);
> if (!mqd)
> return NULL;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> index a9b9882..5dc30f5 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> @@ -108,8 +108,6 @@ static int __update_mqd(struct mqd_manager *mm, void *mqd,
>
> BUG_ON(!mm || !q || !mqd);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> m = get_mqd(mqd);
>
> m->cp_hqd_pq_control = 5 << CP_HQD_PQ_CONTROL__RPTR_BLOCK_SIZE__SHIFT |
> @@ -117,7 +115,7 @@ static int __update_mqd(struct mqd_manager *mm, void *mqd,
> mtype << CP_HQD_PQ_CONTROL__MTYPE__SHIFT;
> m->cp_hqd_pq_control |=
> ffs(q->queue_size / sizeof(unsigned int)) - 1 - 1;
> - pr_debug("kfd: cp_hqd_pq_control 0x%x\n", m->cp_hqd_pq_control);
> + pr_debug("cp_hqd_pq_control 0x%x\n", m->cp_hqd_pq_control);
>
> m->cp_hqd_pq_base_lo = lower_32_bits((uint64_t)q->queue_address >> 8);
> m->cp_hqd_pq_base_hi = upper_32_bits((uint64_t)q->queue_address >> 8);
> @@ -129,7 +127,7 @@ static int __update_mqd(struct mqd_manager *mm, void *mqd,
> 1 << CP_HQD_PQ_DOORBELL_CONTROL__DOORBELL_EN__SHIFT |
> q->doorbell_off <<
> CP_HQD_PQ_DOORBELL_CONTROL__DOORBELL_OFFSET__SHIFT;
> - pr_debug("kfd: cp_hqd_pq_doorbell_control 0x%x\n",
> + pr_debug("cp_hqd_pq_doorbell_control 0x%x\n",
> m->cp_hqd_pq_doorbell_control);
>
> m->cp_hqd_eop_control = atc_bit << CP_HQD_EOP_CONTROL__EOP_ATC__SHIFT |
> @@ -241,8 +239,6 @@ struct mqd_manager *mqd_manager_init_vi(enum KFD_MQD_TYPE type,
> BUG_ON(!dev);
> BUG_ON(type >= KFD_MQD_TYPE_MAX);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> mqd = kzalloc(sizeof(struct mqd_manager), GFP_KERNEL);
> if (!mqd)
> return NULL;
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
> index 99c11a4..31d7d46 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
> @@ -67,7 +67,7 @@ static void pm_calc_rlib_size(struct packet_manager *pm,
> *over_subscription = false;
> if ((process_count > 1) || queue_count > get_queues_num(pm->dqm)) {
> *over_subscription = true;
> - pr_debug("kfd: over subscribed runlist\n");
> + pr_debug("Over subscribed runlist\n");
> }
>
> map_queue_size =
> @@ -85,7 +85,7 @@ static void pm_calc_rlib_size(struct packet_manager *pm,
> if (*over_subscription)
> *rlib_size += sizeof(struct pm4_runlist);
>
> - pr_debug("kfd: runlist ib size %d\n", *rlib_size);
> + pr_debug("runlist ib size %d\n", *rlib_size);
> }
>
> static int pm_allocate_runlist_ib(struct packet_manager *pm,
> @@ -106,7 +106,7 @@ static int pm_allocate_runlist_ib(struct packet_manager *pm,
> &pm->ib_buffer_obj);
>
> if (retval != 0) {
> - pr_err("kfd: failed to allocate runlist IB\n");
> + pr_err("Failed to allocate runlist IB\n");
> return retval;
> }
>
> @@ -152,8 +152,6 @@ static int pm_create_map_process(struct packet_manager *pm, uint32_t *buffer,
>
> packet = (struct pm4_map_process *)buffer;
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> memset(buffer, 0, sizeof(struct pm4_map_process));
>
> packet->header.u32all = build_pm4_header(IT_MAP_PROCESS,
> @@ -189,8 +187,6 @@ static int pm_create_map_queue_vi(struct packet_manager *pm, uint32_t *buffer,
>
> BUG_ON(!pm || !buffer || !q);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> packet = (struct pm4_mes_map_queues *)buffer;
> memset(buffer, 0, sizeof(struct pm4_map_queues));
>
> @@ -223,8 +219,7 @@ static int pm_create_map_queue_vi(struct packet_manager *pm, uint32_t *buffer,
> use_static = false; /* no static queues under SDMA */
> break;
> default:
> - pr_err("kfd: in %s queue type %d\n", __func__,
> - q->properties.type);
> + pr_err("queue type %d\n", q->properties.type);
> BUG();
> break;
> }
> @@ -254,8 +249,6 @@ static int pm_create_map_queue(struct packet_manager *pm, uint32_t *buffer,
>
> BUG_ON(!pm || !buffer || !q);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> packet = (struct pm4_map_queues *)buffer;
> memset(buffer, 0, sizeof(struct pm4_map_queues));
>
> @@ -333,8 +326,7 @@ static int pm_create_runlist_ib(struct packet_manager *pm,
>
> *rl_size_bytes = alloc_size_bytes;
>
> - pr_debug("kfd: In func %s\n", __func__);
> - pr_debug("kfd: building runlist ib process count: %d queues count %d\n",
> + pr_debug("Building runlist ib process count: %d queues count %d\n",
> pm->dqm->processes_count, pm->dqm->queue_count);
>
> /* build the run list ib packet */
> @@ -342,7 +334,7 @@ static int pm_create_runlist_ib(struct packet_manager *pm,
> qpd = cur->qpd;
> /* build map process packet */
> if (proccesses_mapped >= pm->dqm->processes_count) {
> - pr_debug("kfd: not enough space left in runlist IB\n");
> + pr_debug("Not enough space left in runlist IB\n");
> pm_release_ib(pm);
> return -ENOMEM;
> }
> @@ -359,7 +351,7 @@ static int pm_create_runlist_ib(struct packet_manager *pm,
> if (!kq->queue->properties.is_active)
> continue;
>
> - pr_debug("kfd: static_queue, mapping kernel q %d, is debug status %d\n",
> + pr_debug("static_queue, mapping kernel q %d, is debug status %d\n",
> kq->queue->queue, qpd->is_debug);
>
> if (pm->dqm->dev->device_info->asic_family ==
> @@ -385,7 +377,7 @@ static int pm_create_runlist_ib(struct packet_manager *pm,
> if (!q->properties.is_active)
> continue;
>
> - pr_debug("kfd: static_queue, mapping user queue %d, is debug status %d\n",
> + pr_debug("static_queue, mapping user queue %d, is debug status %d\n",
> q->queue, qpd->is_debug);
>
> if (pm->dqm->dev->device_info->asic_family ==
> @@ -409,7 +401,7 @@ static int pm_create_runlist_ib(struct packet_manager *pm,
> }
> }
>
> - pr_debug("kfd: finished map process and queues to runlist\n");
> + pr_debug("Finished map process and queues to runlist\n");
>
> if (is_over_subscription)
> pm_create_runlist(pm, &rl_buffer[rl_wptr], *rl_gpu_addr,
> @@ -453,15 +445,13 @@ int pm_send_set_resources(struct packet_manager *pm,
>
> BUG_ON(!pm || !res);
>
> - pr_debug("kfd: In func %s\n", __func__);
> -
> mutex_lock(&pm->lock);
> pm->priv_queue->ops.acquire_packet_buffer(pm->priv_queue,
> sizeof(*packet) / sizeof(uint32_t),
> (unsigned int **)&packet);
> if (packet == NULL) {
> mutex_unlock(&pm->lock);
> - pr_err("kfd: failed to allocate buffer on kernel queue\n");
> + pr_err("Failed to allocate buffer on kernel queue\n");
> return -ENOMEM;
> }
>
> @@ -504,7 +494,7 @@ int pm_send_runlist(struct packet_manager *pm, struct list_head *dqm_queues)
> if (retval != 0)
> goto fail_create_runlist_ib;
>
> - pr_debug("kfd: runlist IB address: 0x%llX\n", rl_gpu_ib_addr);
> + pr_debug("runlist IB address: 0x%llX\n", rl_gpu_ib_addr);
>
> packet_size_dwords = sizeof(struct pm4_runlist) / sizeof(uint32_t);
> mutex_lock(&pm->lock);
> @@ -595,7 +585,7 @@ int pm_send_unmap_queue(struct packet_manager *pm, enum kfd_queue_type type,
>
> packet = (struct pm4_unmap_queues *)buffer;
> memset(buffer, 0, sizeof(struct pm4_unmap_queues));
> - pr_debug("kfd: static_queue: unmapping queues: mode is %d , reset is %d , type is %d\n",
> + pr_debug("static_queue: unmapping queues: mode is %d , reset is %d , type is %d\n",
> mode, reset, type);
> packet->header.u32all = build_pm4_header(IT_UNMAP_QUEUES,
> sizeof(struct pm4_unmap_queues));
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process.c b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> index a4e4a2d..86032bd 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> @@ -101,7 +101,7 @@ struct kfd_process *kfd_create_process(const struct task_struct *thread)
> /* A prior open of /dev/kfd could have already created the process. */
> process = find_process(thread);
> if (process)
> - pr_debug("kfd: process already found\n");
> + pr_debug("Process already found\n");
>
> if (!process)
> process = create_process(thread);
> @@ -250,7 +250,7 @@ static void kfd_process_notifier_release(struct mmu_notifier *mn,
> kfd_dbgmgr_destroy(pdd->dev->dbgmgr);
>
> if (pdd->reset_wavefronts) {
> - pr_warn("amdkfd: Resetting all wave fronts\n");
> + pr_warn("Resetting all wave fronts\n");
> dbgdev_wave_reset_wavefronts(pdd->dev, p);
> pdd->reset_wavefronts = false;
> }
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> index 32cdf2b..9482a5a 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> @@ -51,15 +51,13 @@ static int find_available_queue_slot(struct process_queue_manager *pqm,
>
> BUG_ON(!pqm || !qid);
>
> - pr_debug("kfd: in %s\n", __func__);
> -
> found = find_first_zero_bit(pqm->queue_slot_bitmap,
> KFD_MAX_NUM_OF_QUEUES_PER_PROCESS);
>
> - pr_debug("kfd: the new slot id %lu\n", found);
> + pr_debug("The new slot id %lu\n", found);
>
> if (found >= KFD_MAX_NUM_OF_QUEUES_PER_PROCESS) {
> - pr_info("amdkfd: Can not open more queues for process with pasid %d\n",
> + pr_info("Cannot open more queues for process with pasid %d\n",
> pqm->process->pasid);
> return -ENOMEM;
> }
> @@ -92,8 +90,6 @@ void pqm_uninit(struct process_queue_manager *pqm)
>
> BUG_ON(!pqm);
>
> - pr_debug("In func %s\n", __func__);
> -
> list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) {
> retval = pqm_destroy_queue(
> pqm,
> @@ -102,7 +98,7 @@ void pqm_uninit(struct process_queue_manager *pqm)
> pqn->kq->queue->properties.queue_id);
>
> if (retval != 0) {
> - pr_err("kfd: failed to destroy queue\n");
> + pr_err("failed to destroy queue\n");
> return;
> }
> }
> @@ -136,7 +132,7 @@ static int create_cp_queue(struct process_queue_manager *pqm,
> (*q)->device = dev;
> (*q)->process = pqm->process;
>
> - pr_debug("kfd: PQM After init queue");
> + pr_debug("PQM After init queue");
>
> return retval;
>
> @@ -210,7 +206,7 @@ int pqm_create_queue(struct process_queue_manager *pqm,
> if ((sched_policy == KFD_SCHED_POLICY_HWS_NO_OVERSUBSCRIPTION) &&
> ((dev->dqm->processes_count >= VMID_PER_DEVICE) ||
> (dev->dqm->queue_count >= get_queues_num(dev->dqm)))) {
> - pr_err("kfd: over-subscription is not allowed in radeon_kfd.sched_policy == 1\n");
> + pr_err("Over-subscription is not allowed in radeon_kfd.sched_policy == 1\n");
> retval = -EPERM;
> goto err_create_queue;
> }
> @@ -243,17 +239,17 @@ int pqm_create_queue(struct process_queue_manager *pqm,
> }
>
> if (retval != 0) {
> - pr_debug("Error dqm create queue\n");
> + pr_err("DQM create queue failed\n");
> goto err_create_queue;
> }
>
> - pr_debug("kfd: PQM After DQM create queue\n");
> + pr_debug("PQM After DQM create queue\n");
>
> list_add(&pqn->process_queue_list, &pqm->queues);
>
> if (q) {
> *properties = q->properties;
> - pr_debug("kfd: PQM done creating queue\n");
> + pr_debug("PQM done creating queue\n");
> print_queue_properties(properties);
> }
>
> @@ -282,11 +278,9 @@ int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid)
> BUG_ON(!pqm);
> retval = 0;
>
> - pr_debug("kfd: In Func %s\n", __func__);
> -
> pqn = get_queue_by_qid(pqm, qid);
> if (pqn == NULL) {
> - pr_err("kfd: queue id does not match any known queue\n");
> + pr_err("Queue id does not match any known queue\n");
> return -EINVAL;
> }
>
> @@ -339,8 +333,7 @@ int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid,
>
> pqn = get_queue_by_qid(pqm, qid);
> if (!pqn) {
> - pr_debug("amdkfd: No queue %d exists for update operation\n",
> - qid);
> + pr_debug("No queue %d exists for update operation\n", qid);
> return -EFAULT;
> }
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_topology.c b/drivers/gpu/drm/amd/amdkfd/kfd_topology.c
> index 0200dae..72d566a 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_topology.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_topology.c
> @@ -666,7 +666,7 @@ static ssize_t node_show(struct kobject *kobj, struct attribute *attr,
> dev->node_props.simd_count);
>
> if (dev->mem_bank_count < dev->node_props.mem_banks_count) {
> - pr_info_once("kfd: mem_banks_count truncated from %d to %d\n",
> + pr_info_once("mem_banks_count truncated from %d to %d\n",
> dev->node_props.mem_banks_count,
> dev->mem_bank_count);
> sysfs_show_32bit_prop(buffer, "mem_banks_count",
> @@ -1147,7 +1147,7 @@ int kfd_topology_add_device(struct kfd_dev *gpu)
>
> gpu_id = kfd_generate_gpu_id(gpu);
>
> - pr_debug("kfd: Adding new GPU (ID: 0x%x) to topology\n", gpu_id);
> + pr_debug("Adding new GPU (ID: 0x%x) to topology\n", gpu_id);
>
> down_write(&topology_lock);
> /*
> @@ -1190,7 +1190,7 @@ int kfd_topology_add_device(struct kfd_dev *gpu)
>
> if (dev->gpu->device_info->asic_family == CHIP_CARRIZO) {
> dev->node_props.capability |= HSA_CAP_DOORBELL_PACKET_TYPE;
> - pr_info("amdkfd: adding doorbell packet type capability\n");
> + pr_info("Adding doorbell packet type capability\n");
> }
>
> res = 0;
> --
> 2.7.4
>
This patch is:
Reviewed-by: Oded Gabbay <oded.gabbay at gmail.com>
More information about the amd-gfx
mailing list