[PATCH 2/2] drm/amdgpu/debugfs: properly handle runtime pm

Quan, Evan Evan.Quan at amd.com
Mon Jan 13 03:36:53 UTC 2020


Reviewed-by: Evan Quan <evan.quan at amd.com>

> -----Original Message-----
> From: amd-gfx <amd-gfx-bounces at lists.freedesktop.org> On Behalf Of Alex
> Deucher
> Sent: Saturday, January 11, 2020 7:45 AM
> To: amd-gfx at lists.freedesktop.org
> Cc: Deucher, Alexander <Alexander.Deucher at amd.com>
> Subject: [PATCH 2/2] drm/amdgpu/debugfs: properly handle runtime pm
> 
> If driver debugfs files are accessed, power up the GPU when necessary.
> 
> Signed-off-by: Alex Deucher <alexander.deucher at amd.com>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c | 133
> ++++++++++++++++++--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c   |   8 ++
>  2 files changed, 134 insertions(+), 7 deletions(-)
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> index 63343bb43049..f24ed9a1a3e5 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
> @@ -26,6 +26,7 @@
>  #include <linux/kthread.h>
>  #include <linux/pci.h>
>  #include <linux/uaccess.h>
> +#include <linux/pm_runtime.h>
> 
>  #include <drm/drm_debugfs.h>
> 
> @@ -144,10 +145,17 @@ static int  amdgpu_debugfs_process_reg_op(bool
> read, struct file *f,
> 
>  	*pos &= (1UL << 22) - 1;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	if (use_bank) {
>  		if ((sh_bank != 0xFFFFFFFF && sh_bank >= adev-
> >gfx.config.max_sh_per_se) ||
> -		    (se_bank != 0xFFFFFFFF && se_bank >= adev-
> >gfx.config.max_shader_engines))
> +		    (se_bank != 0xFFFFFFFF && se_bank >= adev-
> >gfx.config.max_shader_engines)) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return -EINVAL;
> +		}
>  		mutex_lock(&adev->grbm_idx_mutex);
>  		amdgpu_gfx_select_se_sh(adev, se_bank,
>  					sh_bank, instance_bank);
> @@ -193,6 +201,9 @@ static int  amdgpu_debugfs_process_reg_op(bool read,
> struct file *f,
>  	if (pm_pg_lock)
>  		mutex_unlock(&adev->pm.mutex);
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -237,13 +248,20 @@ static ssize_t
> amdgpu_debugfs_regs_pcie_read(struct file *f, char __user *buf,
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		value = RREG32_PCIE(*pos >> 2);
>  		r = put_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		result += 4;
>  		buf += 4;
> @@ -251,6 +269,9 @@ static ssize_t amdgpu_debugfs_regs_pcie_read(struct
> file *f, char __user *buf,
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -276,12 +297,19 @@ static ssize_t
> amdgpu_debugfs_regs_pcie_write(struct file *f, const char __user
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		r = get_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		WREG32_PCIE(*pos >> 2, value);
> 
> @@ -291,6 +319,9 @@ static ssize_t amdgpu_debugfs_regs_pcie_write(struct
> file *f, const char __user
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -316,13 +347,20 @@ static ssize_t
> amdgpu_debugfs_regs_didt_read(struct file *f, char __user *buf,
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		value = RREG32_DIDT(*pos >> 2);
>  		r = put_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		result += 4;
>  		buf += 4;
> @@ -330,6 +368,9 @@ static ssize_t amdgpu_debugfs_regs_didt_read(struct
> file *f, char __user *buf,
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -355,12 +396,19 @@ static ssize_t
> amdgpu_debugfs_regs_didt_write(struct file *f, const char __user
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		r = get_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		WREG32_DIDT(*pos >> 2, value);
> 
> @@ -370,6 +418,9 @@ static ssize_t amdgpu_debugfs_regs_didt_write(struct
> file *f, const char __user
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -395,13 +446,20 @@ static ssize_t
> amdgpu_debugfs_regs_smc_read(struct file *f, char __user *buf,
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		value = RREG32_SMC(*pos);
>  		r = put_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		result += 4;
>  		buf += 4;
> @@ -409,6 +467,9 @@ static ssize_t amdgpu_debugfs_regs_smc_read(struct
> file *f, char __user *buf,
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -434,12 +495,19 @@ static ssize_t
> amdgpu_debugfs_regs_smc_write(struct file *f, const char __user *
>  	if (size & 0x3 || *pos & 0x3)
>  		return -EINVAL;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	while (size) {
>  		uint32_t value;
> 
>  		r = get_user(value, (uint32_t *)buf);
> -		if (r)
> +		if (r) {
> +			pm_runtime_mark_last_busy(adev->ddev->dev);
> +			pm_runtime_put_autosuspend(adev->ddev->dev);
>  			return r;
> +		}
> 
>  		WREG32_SMC(*pos, value);
> 
> @@ -449,6 +517,9 @@ static ssize_t amdgpu_debugfs_regs_smc_write(struct
> file *f, const char __user *
>  		size -= 4;
>  	}
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	return result;
>  }
> 
> @@ -572,7 +643,16 @@ static ssize_t amdgpu_debugfs_sensor_read(struct file
> *f, char __user *buf,
>  	idx = *pos >> 2;
> 
>  	valuesize = sizeof(values);
> +
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	r = amdgpu_dpm_read_sensor(adev, idx, &values[0], &valuesize);
> +
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	if (r)
>  		return r;
> 
> @@ -633,6 +713,10 @@ static ssize_t amdgpu_debugfs_wave_read(struct file
> *f, char __user *buf,
>  	wave = (*pos & GENMASK_ULL(36, 31)) >> 31;
>  	simd = (*pos & GENMASK_ULL(44, 37)) >> 37;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	/* switch to the specific se/sh/cu */
>  	mutex_lock(&adev->grbm_idx_mutex);
>  	amdgpu_gfx_select_se_sh(adev, se, sh, cu); @@ -644,6 +728,9 @@
> static ssize_t amdgpu_debugfs_wave_read(struct file *f, char __user *buf,
>  	amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF);
>  	mutex_unlock(&adev->grbm_idx_mutex);
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	if (!x)
>  		return -EINVAL;
> 
> @@ -711,6 +798,10 @@ static ssize_t amdgpu_debugfs_gpr_read(struct file *f,
> char __user *buf,
>  	if (!data)
>  		return -ENOMEM;
> 
> +	r = pm_runtime_get_sync(adev->ddev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	/* switch to the specific se/sh/cu */
>  	mutex_lock(&adev->grbm_idx_mutex);
>  	amdgpu_gfx_select_se_sh(adev, se, sh, cu); @@ -726,6 +817,9 @@
> static ssize_t amdgpu_debugfs_gpr_read(struct file *f, char __user *buf,
>  	amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF);
>  	mutex_unlock(&adev->grbm_idx_mutex);
> 
> +	pm_runtime_mark_last_busy(adev->ddev->dev);
> +	pm_runtime_put_autosuspend(adev->ddev->dev);
> +
>  	while (size) {
>  		uint32_t value;
> 
> @@ -859,6 +953,10 @@ static int amdgpu_debugfs_test_ib(struct seq_file *m,
> void *data)
>  	struct amdgpu_device *adev = dev->dev_private;
>  	int r = 0, i;
> 
> +	r = pm_runtime_get_sync(dev->dev);
> +	if (r < 0)
> +		return r;
> +
>  	/* Avoid accidently unparking the sched thread during GPU reset */
>  	mutex_lock(&adev->lock_reset);
> 
> @@ -889,6 +987,9 @@ static int amdgpu_debugfs_test_ib(struct seq_file *m,
> void *data)
> 
>  	mutex_unlock(&adev->lock_reset);
> 
> +	pm_runtime_mark_last_busy(dev->dev);
> +	pm_runtime_put_autosuspend(dev->dev);
> +
>  	return 0;
>  }
> 
> @@ -907,8 +1008,17 @@ static int amdgpu_debugfs_evict_vram(struct
> seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *)m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct amdgpu_device *adev = dev->dev_private;
> +	int r;
> +
> +	r = pm_runtime_get_sync(dev->dev);
> +	if (r < 0)
> +		return r;
> 
>  	seq_printf(m, "(%d)\n", amdgpu_bo_evict_vram(adev));
> +
> +	pm_runtime_mark_last_busy(dev->dev);
> +	pm_runtime_put_autosuspend(dev->dev);
> +
>  	return 0;
>  }
> 
> @@ -917,8 +1027,17 @@ static int amdgpu_debugfs_evict_gtt(struct seq_file
> *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *)m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct amdgpu_device *adev = dev->dev_private;
> +	int r;
> +
> +	r = pm_runtime_get_sync(dev->dev);
> +	if (r < 0)
> +		return r;
> 
>  	seq_printf(m, "(%d)\n", ttm_bo_evict_mm(&adev->mman.bdev,
> TTM_PL_TT));
> +
> +	pm_runtime_mark_last_busy(dev->dev);
> +	pm_runtime_put_autosuspend(dev->dev);
> +
>  	return 0;
>  }
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> index e9efee04ca23..3c01252b1e0e 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
> @@ -741,10 +741,18 @@ static int amdgpu_debugfs_gpu_recover(struct
> seq_file *m, void *data)
>  	struct drm_info_node *node = (struct drm_info_node *) m->private;
>  	struct drm_device *dev = node->minor->dev;
>  	struct amdgpu_device *adev = dev->dev_private;
> +	int r;
> +
> +	r = pm_runtime_get_sync(dev->dev);
> +	if (r < 0)
> +		return 0;
> 
>  	seq_printf(m, "gpu recover\n");
>  	amdgpu_device_gpu_recover(adev, NULL);
> 
> +	pm_runtime_mark_last_busy(dev->dev);
> +	pm_runtime_put_autosuspend(dev->dev);
> +
>  	return 0;
>  }
> 
> --
> 2.24.1
> 
> _______________________________________________
> amd-gfx mailing list
> amd-gfx at lists.freedesktop.org
> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.free
> desktop.org%2Fmailman%2Flistinfo%2Famd-
> gfx&data=02%7C01%7Cevan.quan%40amd.com%7C8e6ecdab282646758a
> de08d796272aa0%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637
> 142967348327101&sdata=KaUJb0eNiSNov6amPZ20oxPm6howvbmU093i2
> EAprO0%3D&reserved=0


More information about the amd-gfx mailing list