[PATCH v3] drm/amdkfd: Track SDMA utilization per process
Felix Kuehling
felix.kuehling at amd.com
Tue May 26 20:07:32 UTC 2020
Am 2020-05-26 um 1:26 p.m. schrieb Mukul Joshi:
> Track SDMA usage on a per process basis and report it through sysfs.
> The value in the sysfs file indicates the amount of time SDMA has
> been in-use by this process since the creation of the process.
> This value is in microsecond granularity.
>
> v2:
> - Remove unnecessary checks for pdd is kfd_procfs_show().
> - Make counter part of the kfd_sdma_activity_handler_workarea
> structure.
>
> v3:
> - Remove READ_ONCE/WRITE_ONCE while updating acitivty
Typo: activity
> counter.
> - Add updation of past acitivt counter under dqm_lock.
Typo: activity
Other than that, the patch is
Reviewed-by: Felix Kuehling <Felix.Kuehling at amd.com>
>
> Signed-off-by: Mukul Joshi <mukul.joshi at amd.com>
> ---
> .../drm/amd/amdkfd/kfd_device_queue_manager.c | 57 ++++++++
> .../drm/amd/amdkfd/kfd_device_queue_manager.h | 2 +
> drivers/gpu/drm/amd/amdkfd/kfd_priv.h | 16 +-
> drivers/gpu/drm/amd/amdkfd/kfd_process.c | 137 ++++++++++++++++--
> 4 files changed, 198 insertions(+), 14 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> index e9c4867abeff..6293017bd5bf 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> @@ -153,6 +153,52 @@ void decrement_queue_count(struct device_queue_manager *dqm,
> dqm->active_cp_queue_count--;
> }
>
> +int read_sdma_queue_counter(struct queue *q, uint64_t *val)
> +{
> + int ret;
> + uint64_t tmp = 0;
> +
> + if (!q || !val)
> + return -EINVAL;
> + /*
> + * SDMA activity counter is stored at queue's RPTR + 0x8 location.
> + */
> + if (!access_ok((const void __user *)((uint64_t)q->properties.read_ptr +
> + sizeof(uint64_t)), sizeof(uint64_t))) {
> + pr_err("Can't access sdma queue activity counter\n");
> + return -EFAULT;
> + }
> +
> + ret = get_user(tmp, (uint64_t *)((uint64_t)(q->properties.read_ptr) +
> + sizeof(uint64_t)));
> + if (!ret) {
> + *val = tmp;
> + }
> +
> + return ret;
> +}
> +
> +static int update_sdma_queue_past_activity_stats(struct kfd_process_device *pdd,
> + struct queue *q)
> +{
> + int ret;
> + uint64_t val = 0;
> +
> + if (!pdd)
> + return -ENODEV;
> +
> + ret = read_sdma_queue_counter(q, &val);
> + if (ret) {
> + pr_err("Failed to read SDMA queue counter for queue: %d\n",
> + q->properties.queue_id);
> + return ret;
> + }
> +
> + pdd->sdma_past_activity_counter += val;
> +
> + return ret;
> +}
> +
> static int allocate_doorbell(struct qcm_process_device *qpd, struct queue *q)
> {
> struct kfd_dev *dev = qpd->dqm->dev;
> @@ -487,6 +533,12 @@ static int destroy_queue_nocpsch_locked(struct device_queue_manager *dqm,
> if (retval == -ETIME)
> qpd->reset_wavefronts = true;
>
> + /* Get the SDMA queue stats */
> + if ((q->properties.type == KFD_QUEUE_TYPE_SDMA) ||
> + (q->properties.type == KFD_QUEUE_TYPE_SDMA_XGMI)) {
> + update_sdma_queue_past_activity_stats(qpd_to_pdd(qpd), q);
> + }
> +
> mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj);
>
> list_del(&q->list);
> @@ -1468,6 +1520,11 @@ static int destroy_queue_cpsch(struct device_queue_manager *dqm,
> }
> }
>
> + /* Get the SDMA queue stats */
> + if ((q->properties.type == KFD_QUEUE_TYPE_SDMA) ||
> + (q->properties.type == KFD_QUEUE_TYPE_SDMA_XGMI)) {
> + update_sdma_queue_past_activity_stats(qpd_to_pdd(qpd), q);
> + }
> /*
> * Unconditionally decrement this counter, regardless of the queue's
> * type
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
> index 4afa015c69b1..894bcf877f9e 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
> @@ -251,4 +251,6 @@ static inline void dqm_unlock(struct device_queue_manager *dqm)
> mutex_unlock(&dqm->lock_hidden);
> }
>
> +int read_sdma_queue_counter(struct queue *q, uint64_t *val);
> +
> #endif /* KFD_DEVICE_QUEUE_MANAGER_H_ */
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> index f70f789c3cb3..fae139b77c0a 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> @@ -633,7 +633,14 @@ enum kfd_pdd_bound {
> PDD_BOUND_SUSPENDED,
> };
>
> -#define MAX_VRAM_FILENAME_LEN 11
> +#define MAX_SYSFS_FILENAME_LEN 11
> +
> +/*
> + * SDMA counter runs at 100MHz frequency.
> + * We display SDMA activity in microsecond granularity in sysfs.
> + * As a result, the divisor is 100.
> + */
> +#define SDMA_ACTIVITY_DIVISOR 100
>
> /* Data that is per-process-per device. */
> struct kfd_process_device {
> @@ -681,7 +688,12 @@ struct kfd_process_device {
> /* VRAM usage */
> uint64_t vram_usage;
> struct attribute attr_vram;
> - char vram_filename[MAX_VRAM_FILENAME_LEN];
> + char vram_filename[MAX_SYSFS_FILENAME_LEN];
> +
> + /* SDMA activity tracking */
> + uint64_t sdma_past_activity_counter;
> + struct attribute attr_sdma;
> + char sdma_filename[MAX_SYSFS_FILENAME_LEN];
> };
>
> #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd)
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process.c b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> index d27221ddcdeb..db010c5da144 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process.c
> @@ -25,6 +25,7 @@
> #include <linux/sched.h>
> #include <linux/sched/mm.h>
> #include <linux/sched/task.h>
> +#include <linux/mmu_context.h>
> #include <linux/slab.h>
> #include <linux/amd-iommu.h>
> #include <linux/notifier.h>
> @@ -76,6 +77,74 @@ struct kfd_procfs_tree {
>
> static struct kfd_procfs_tree procfs;
>
> +/*
> + * Structure for SDMA activity tracking
> + */
> +struct kfd_sdma_activity_handler_workarea {
> + struct work_struct sdma_activity_work;
> + struct kfd_process_device *pdd;
> + uint64_t sdma_activity_counter;
> +};
> +
> +static void kfd_sdma_activity_worker(struct work_struct *work)
> +{
> + struct kfd_sdma_activity_handler_workarea *workarea;
> + struct kfd_process_device *pdd;
> + uint64_t val;
> + struct mm_struct *mm;
> + struct queue *q;
> + struct qcm_process_device *qpd;
> + struct device_queue_manager *dqm;
> + int ret = 0;
> +
> + workarea = container_of(work, struct kfd_sdma_activity_handler_workarea,
> + sdma_activity_work);
> + if (!workarea)
> + return;
> +
> + pdd = workarea->pdd;
> + dqm = pdd->dev->dqm;
> + qpd = &pdd->qpd;
> +
> + if (!pdd || !dqm || !qpd)
> + return;
> +
> + mm = get_task_mm(pdd->process->lead_thread);
> + if (!mm) {
> + return;
> + }
> +
> + use_mm(mm);
> +
> + dqm_lock(dqm);
> +
> + /*
> + * Total SDMA activity is current SDMA activity + past SDMA activity
> + */
> + workarea->sdma_activity_counter = pdd->sdma_past_activity_counter;
> +
> + /*
> + * Get the current activity counters for all active SDMA queues
> + */
> + list_for_each_entry(q, &qpd->queues_list, list) {
> + if ((q->properties.type == KFD_QUEUE_TYPE_SDMA) ||
> + (q->properties.type == KFD_QUEUE_TYPE_SDMA_XGMI)) {
> + val = 0;
> + ret = read_sdma_queue_counter(q, &val);
> + if (ret)
> + pr_debug("Failed to read SDMA queue active "
> + "counter for queue id: %d",
> + q->properties.queue_id);
> + else
> + workarea->sdma_activity_counter += val;
> + }
> + }
> +
> + dqm_unlock(dqm);
> + unuse_mm(mm);
> + mmput(mm);
> +}
> +
> static ssize_t kfd_procfs_show(struct kobject *kobj, struct attribute *attr,
> char *buffer)
> {
> @@ -87,8 +156,24 @@ static ssize_t kfd_procfs_show(struct kobject *kobj, struct attribute *attr,
> } else if (strncmp(attr->name, "vram_", 5) == 0) {
> struct kfd_process_device *pdd = container_of(attr, struct kfd_process_device,
> attr_vram);
> - if (pdd)
> - return snprintf(buffer, PAGE_SIZE, "%llu\n", READ_ONCE(pdd->vram_usage));
> + return snprintf(buffer, PAGE_SIZE, "%llu\n", READ_ONCE(pdd->vram_usage));
> + } else if (strncmp(attr->name, "sdma_", 5) == 0) {
> + struct kfd_process_device *pdd = container_of(attr, struct kfd_process_device,
> + attr_sdma);
> + struct kfd_sdma_activity_handler_workarea sdma_activity_work_handler;
> +
> + INIT_WORK(&sdma_activity_work_handler.sdma_activity_work,
> + kfd_sdma_activity_worker);
> +
> + sdma_activity_work_handler.pdd = pdd;
> +
> + schedule_work(&sdma_activity_work_handler.sdma_activity_work);
> +
> + flush_work(&sdma_activity_work_handler.sdma_activity_work);
> +
> + return snprintf(buffer, PAGE_SIZE, "%llu\n",
> + (sdma_activity_work_handler.sdma_activity_counter)/
> + SDMA_ACTIVITY_DIVISOR);
> } else {
> pr_err("Invalid attribute");
> return -EINVAL;
> @@ -210,7 +295,24 @@ int kfd_procfs_add_queue(struct queue *q)
> return 0;
> }
>
> -int kfd_procfs_add_vram_usage(struct kfd_process *p)
> +static int kfd_sysfs_create_file(struct kfd_process *p, struct attribute *attr,
> + char *name)
> +{
> + int ret = 0;
> +
> + if (!p || !attr || !name)
> + return -EINVAL;
> +
> + attr->name = name;
> + attr->mode = KFD_SYSFS_FILE_MODE;
> + sysfs_attr_init(attr);
> +
> + ret = sysfs_create_file(p->kobj, attr);
> +
> + return ret;
> +}
> +
> +int kfd_procfs_add_sysfs_files(struct kfd_process *p)
> {
> int ret = 0;
> struct kfd_process_device *pdd;
> @@ -221,17 +323,25 @@ int kfd_procfs_add_vram_usage(struct kfd_process *p)
> if (!p->kobj)
> return -EFAULT;
>
> - /* Create proc/<pid>/vram_<gpuid> file for each GPU */
> + /*
> + * Create sysfs files for each GPU:
> + * - proc/<pid>/vram_<gpuid>
> + * - proc/<pid>/sdma_<gpuid>
> + */
> list_for_each_entry(pdd, &p->per_device_data, per_device_list) {
> - snprintf(pdd->vram_filename, MAX_VRAM_FILENAME_LEN, "vram_%u",
> + snprintf(pdd->vram_filename, MAX_SYSFS_FILENAME_LEN, "vram_%u",
> pdd->dev->id);
> - pdd->attr_vram.name = pdd->vram_filename;
> - pdd->attr_vram.mode = KFD_SYSFS_FILE_MODE;
> - sysfs_attr_init(&pdd->attr_vram);
> - ret = sysfs_create_file(p->kobj, &pdd->attr_vram);
> + ret = kfd_sysfs_create_file(p, &pdd->attr_vram, pdd->vram_filename);
> if (ret)
> pr_warn("Creating vram usage for gpu id %d failed",
> (int)pdd->dev->id);
> +
> + snprintf(pdd->sdma_filename, MAX_SYSFS_FILENAME_LEN, "sdma_%u",
> + pdd->dev->id);
> + ret = kfd_sysfs_create_file(p, &pdd->attr_sdma, pdd->sdma_filename);
> + if (ret)
> + pr_warn("Creating sdma usage for gpu id %d failed",
> + (int)pdd->dev->id);
> }
>
> return ret;
> @@ -444,9 +554,9 @@ struct kfd_process *kfd_create_process(struct file *filep)
> if (!process->kobj_queues)
> pr_warn("Creating KFD proc/queues folder failed");
>
> - ret = kfd_procfs_add_vram_usage(process);
> + ret = kfd_procfs_add_sysfs_files(process);
> if (ret)
> - pr_warn("Creating vram usage file for pid %d failed",
> + pr_warn("Creating sysfs usage file for pid %d failed",
> (int)process->lead_thread->pid);
> }
> out:
> @@ -597,8 +707,10 @@ static void kfd_process_wq_release(struct work_struct *work)
> kobject_put(p->kobj_queues);
> p->kobj_queues = NULL;
>
> - list_for_each_entry(pdd, &p->per_device_data, per_device_list)
> + list_for_each_entry(pdd, &p->per_device_data, per_device_list) {
> sysfs_remove_file(p->kobj, &pdd->attr_vram);
> + sysfs_remove_file(p->kobj, &pdd->attr_sdma);
> + }
>
> kobject_del(p->kobj);
> kobject_put(p->kobj);
> @@ -906,6 +1018,7 @@ struct kfd_process_device *kfd_create_process_device_data(struct kfd_dev *dev,
> pdd->already_dequeued = false;
> pdd->runtime_inuse = false;
> pdd->vram_usage = 0;
> + pdd->sdma_past_activity_counter = 0;
> list_add(&pdd->per_device_list, &p->per_device_data);
>
> /* Init idr used for memory handle translation */
More information about the amd-gfx
mailing list