[Intel-gfx] [RFC 8/8] drm/i915: Expose per-engine client busyness
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Thu Dec 19 18:00:19 UTC 2019
From: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
Expose per-client and per-engine busyness under the previously added sysfs
client root.
The new files are one per-engine instance and located under the 'busy'
directory. Each contains a monotonically increasing nano-second resolution
times each client's jobs were executing on the GPU.
This enables userspace to create a top-like tool for GPU utilization:
==========================================================================
intel-gpu-top - 935/ 935 MHz; 0% RC6; 14.73 Watts; 1097 irqs/s
IMC reads: 1401 MiB/s
IMC writes: 4 MiB/s
ENGINE BUSY MI_SEMA MI_WAIT
Render/3D/0 63.73% |███████████████████ | 3% 0%
Blitter/0 9.53% |██▊ | 6% 0%
Video/0 39.32% |███████████▊ | 16% 0%
Video/1 15.62% |████▋ | 0% 0%
VideoEnhance/0 0.00% | | 0% 0%
PID NAME RCS BCS VCS VECS
4084 gem_wsim |█████▌ ||█ || || |
4086 gem_wsim |█▌ || ||███ || |
==========================================================================
v2: Use intel_context_engine_get_busy_time.
v3: New directory structure.
v4: Rebase.
v5: sysfs_attr_init.
v6: Small tidy in i915_gem_add_client.
v7: Rebase to be engine class based.
v8:
* Always enable stats.
* Walk all client contexts.
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
---
drivers/gpu/drm/i915/i915_drv.h | 9 ++++
drivers/gpu/drm/i915/i915_gem.c | 89 ++++++++++++++++++++++++++++++++-
2 files changed, 97 insertions(+), 1 deletion(-)
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index 8ffd638a071f..8cba3cfb5910 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -189,6 +189,13 @@ struct drm_i915_private;
struct i915_mm_struct;
struct i915_mmu_object;
struct i915_drm_clients;
+struct i915_drm_client;
+
+struct i915_engine_busy_attribute {
+ struct device_attribute attr;
+ struct i915_drm_client *client;
+ unsigned int engine_class;
+};
struct drm_i915_file_private {
struct kref kref;
@@ -238,9 +245,11 @@ struct drm_i915_file_private {
struct list_head ctx_list;
struct kobject *root;
+ struct kobject *busy_root;
struct {
struct device_attribute pid;
struct device_attribute name;
+ struct i915_engine_busy_attribute busy[MAX_ENGINE_CLASS];
} attr;
} client;
};
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 564e21902dff..98cee37931f6 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -1546,13 +1546,56 @@ show_client_pid(struct device *kdev, struct device_attribute *attr, char *buf)
return snprintf(buf, PAGE_SIZE, "-");
}
+static u64 busy_add(struct i915_gem_context *ctx, unsigned int engine_class)
+{
+ struct i915_gem_engines *engines = rcu_dereference(ctx->engines);
+ struct i915_gem_engines_iter it;
+ struct intel_context *ce;
+ u64 total = 0;
+
+ for_each_gem_engine(ce, engines, it) {
+ if (ce->engine->uabi_class == engine_class)
+ total += ktime_to_ns(intel_context_get_busy_time(ce));
+ }
+
+ return total;
+}
+
+static ssize_t
+show_client_busy(struct device *kdev, struct device_attribute *attr, char *buf)
+{
+ struct i915_engine_busy_attribute *i915_attr =
+ container_of(attr, typeof(*i915_attr), attr);
+ struct list_head *list = &i915_attr->client->ctx_list;
+ unsigned int engine_class = i915_attr->engine_class;
+ struct i915_gem_context *ctx;
+ u64 total = 0;
+
+ rcu_read_lock();
+ list_for_each_entry_rcu(ctx, list, client_link)
+ total += busy_add(ctx, engine_class);
+ rcu_read_unlock();
+
+ return snprintf(buf, PAGE_SIZE, "%llu\n", total);
+}
+
+static const char *uabi_class_names[] = {
+ [I915_ENGINE_CLASS_RENDER] = "0",
+ [I915_ENGINE_CLASS_COPY] = "1",
+ [I915_ENGINE_CLASS_VIDEO] = "2",
+ [I915_ENGINE_CLASS_VIDEO_ENHANCE] = "3",
+};
+
int
__i915_gem_register_client(struct i915_drm_clients *clients,
struct i915_drm_client *client,
struct task_struct *task)
{
+ struct drm_i915_private *i915 =
+ container_of(clients, typeof(*i915), clients);
struct device_attribute *attr;
- int ret = -ENOMEM;
+ struct intel_engine_cs *engine;
+ int i, ret = -ENOMEM;
char idstr[32];
if (!clients->root)
@@ -1587,10 +1630,47 @@ __i915_gem_register_client(struct i915_drm_clients *clients,
if (ret)
goto err_attr;
+ if (i915->caps.scheduler & I915_SCHEDULER_CAP_ENGINE_BUSY_STATS) {
+ client->busy_root =
+ kobject_create_and_add("busy", client->root);
+ if (!client->busy_root)
+ goto err_attr;
+ }
+
+ for (i = 0;
+ client->busy_root && i < ARRAY_SIZE(uabi_class_names);
+ i++) {
+ struct i915_engine_busy_attribute *i915_attr =
+ &client->attr.busy[i];
+
+ i915_attr->client = client;
+ i915_attr->engine_class = i;
+
+ attr = &i915_attr->attr;
+
+ sysfs_attr_init(&attr->attr);
+
+ attr->attr.name = uabi_class_names[i];
+ attr->attr.mode = 0444;
+ attr->show = show_client_busy;
+
+ ret = sysfs_create_file(client->busy_root,
+ (struct attribute *)attr);
+ if (ret)
+ goto err_busy;
+ }
+
client->pid = get_task_pid(task, PIDTYPE_PID);
+ if (client->busy_root) {
+ for_each_uabi_engine(engine, i915)
+ WARN_ON_ONCE(intel_enable_engine_stats(engine));
+ }
+
return 0;
+err_busy:
+ kobject_put(client->busy_root);
err_attr:
kobject_put(client->root);
err_client:
@@ -1603,10 +1683,17 @@ void __i915_gem_unregister_client(struct i915_drm_client *client)
{
struct drm_i915_file_private *fpriv =
container_of(client, typeof(*fpriv), client);
+ struct intel_engine_cs *engine;
if (!client->name)
return; /* intel_fbdev_init registers a client before sysfs */
+ if (client->busy_root) {
+ for_each_uabi_engine(engine, fpriv->i915)
+ intel_disable_engine_stats(engine);
+ }
+
+ kobject_put(fetch_and_zero(&client->busy_root));
kobject_put(fetch_and_zero(&client->root));
put_pid(fetch_and_zero(&client->pid));
kfree(fetch_and_zero(&client->name));
--
2.20.1
More information about the Intel-gfx
mailing list