[PATCH] drm/amdgpu: add flag for high priority contexts
Alex Deucher
alexdeucher at gmail.com
Tue Jan 3 23:02:48 UTC 2017
On Tue, Jan 3, 2017 at 6:00 PM, Andres Rodriguez <andresx7 at gmail.com> wrote:
> I was thinking of that originally, but the allocation context already has a
> flags variable which allows us to preserve the IOCTL ABI.
>
> We could reserve a few bits of the flags for a priority level instead if
> that sounds good?
We can also use _pad for something else.
Alex
>
> Regards,
> Andres
>
> On Tue, Jan 3, 2017 at 5:59 PM, Alex Deucher <alexdeucher at gmail.com> wrote:
>>
>> On Tue, Jan 3, 2017 at 5:54 PM, Andres Rodriguez <andresx7 at gmail.com>
>> wrote:
>> > Add a new context creation flag, AMDGPU_CTX_FLAG_HIGHPRIORITY. This flag
>> > results in the allocated context receiving a higher scheduler priority
>> > that other contexts system-wide.
>> >
>> > Signed-off-by: Andres Rodriguez <andresx7 at gmail.com>
>> > ---
>> > drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c | 24
>> > ++++++++++++++++++------
>> > drivers/gpu/drm/amd/scheduler/gpu_scheduler.h | 1 +
>> > include/uapi/drm/amdgpu_drm.h | 3 ++-
>> > 3 files changed, 21 insertions(+), 7 deletions(-)
>> >
>> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
>> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
>> > index 400c66b..ce470e2 100644
>> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
>> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
>> > @@ -25,11 +25,17 @@
>> > #include <drm/drmP.h>
>> > #include "amdgpu.h"
>> >
>> > -static int amdgpu_ctx_init(struct amdgpu_device *adev, struct
>> > amdgpu_ctx *ctx)
>> > +static int amdgpu_ctx_init(struct amdgpu_device *adev, int priority,
>> > struct amdgpu_ctx *ctx)
>> > {
>> > unsigned i, j;
>> > int r;
>> >
>> > + if (priority < 0 || priority >= AMD_SCHED_MAX_PRIORITY)
>> > + return -EINVAL;
>> > +
>> > + if (priority == AMD_SCHED_PRIORITY_HIGH &&
>> > !capable(CAP_SYS_ADMIN))
>> > + return -EACCES;
>> > +
>> > memset(ctx, 0, sizeof(*ctx));
>> > ctx->adev = adev;
>> > kref_init(&ctx->refcount);
>> > @@ -51,7 +57,7 @@ static int amdgpu_ctx_init(struct amdgpu_device *adev,
>> > struct amdgpu_ctx *ctx)
>> > struct amdgpu_ring *ring = adev->rings[i];
>> > struct amd_sched_rq *rq;
>> >
>> > - rq = &ring->sched.sched_rq[AMD_SCHED_PRIORITY_NORMAL];
>> > + rq = &ring->sched.sched_rq[priority];
>> > r = amd_sched_entity_init(&ring->sched,
>> > &ctx->rings[i].entity,
>> > rq, amdgpu_sched_jobs);
>> > if (r)
>> > @@ -90,11 +96,15 @@ static void amdgpu_ctx_fini(struct amdgpu_ctx *ctx)
>> >
>> > static int amdgpu_ctx_alloc(struct amdgpu_device *adev,
>> > struct amdgpu_fpriv *fpriv,
>> > + int flags,
>> > uint32_t *id)
>> > {
>> > struct amdgpu_ctx_mgr *mgr = &fpriv->ctx_mgr;
>> > struct amdgpu_ctx *ctx;
>> > - int r;
>> > + int r, priority;
>> > +
>> > + priority = flags & AMDGPU_CTX_FLAG_HIGHPRIORITY ?
>> > + AMD_SCHED_PRIORITY_HIGH : AMD_SCHED_PRIORITY_NORMAL;
>> >
>> > ctx = kmalloc(sizeof(*ctx), GFP_KERNEL);
>> > if (!ctx)
>> > @@ -107,8 +117,9 @@ static int amdgpu_ctx_alloc(struct amdgpu_device
>> > *adev,
>> > kfree(ctx);
>> > return r;
>> > }
>> > +
>> > *id = (uint32_t)r;
>> > - r = amdgpu_ctx_init(adev, ctx);
>> > + r = amdgpu_ctx_init(adev, priority, ctx);
>> > if (r) {
>> > idr_remove(&mgr->ctx_handles, *id);
>> > *id = 0;
>> > @@ -186,7 +197,7 @@ int amdgpu_ctx_ioctl(struct drm_device *dev, void
>> > *data,
>> > struct drm_file *filp)
>> > {
>> > int r;
>> > - uint32_t id;
>> > + uint32_t id, flags;
>> >
>> > union drm_amdgpu_ctx *args = data;
>> > struct amdgpu_device *adev = dev->dev_private;
>> > @@ -194,10 +205,11 @@ int amdgpu_ctx_ioctl(struct drm_device *dev, void
>> > *data,
>> >
>> > r = 0;
>> > id = args->in.ctx_id;
>> > + flags = args->in.flags;
>> >
>> > switch (args->in.op) {
>> > case AMDGPU_CTX_OP_ALLOC_CTX:
>> > - r = amdgpu_ctx_alloc(adev, fpriv, &id);
>> > + r = amdgpu_ctx_alloc(adev, fpriv, flags, &id);
>> > args->out.alloc.ctx_id = id;
>> > break;
>> > case AMDGPU_CTX_OP_FREE_CTX:
>> > diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
>> > b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
>> > index d8dc681..2e458de 100644
>> > --- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
>> > +++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
>> > @@ -108,6 +108,7 @@ struct amd_sched_backend_ops {
>> >
>> > enum amd_sched_priority {
>> > AMD_SCHED_PRIORITY_KERNEL = 0,
>> > + AMD_SCHED_PRIORITY_HIGH,
>> > AMD_SCHED_PRIORITY_NORMAL,
>> > AMD_SCHED_MAX_PRIORITY
>> > };
>> > diff --git a/include/uapi/drm/amdgpu_drm.h
>> > b/include/uapi/drm/amdgpu_drm.h
>> > index 3961836..702332e 100644
>> > --- a/include/uapi/drm/amdgpu_drm.h
>> > +++ b/include/uapi/drm/amdgpu_drm.h
>> > @@ -160,10 +160,11 @@ union drm_amdgpu_bo_list {
>> > /* unknown cause */
>> > #define AMDGPU_CTX_UNKNOWN_RESET 3
>> >
>> > +#define AMDGPU_CTX_FLAG_HIGHPRIORITY (1 << 0)
>>
>> Would it be better to expose a priority level rather than a single
>> flag? If we want to expose more than just normal/high in the future
>> for example.
>>
>> Alex
>>
>> > struct drm_amdgpu_ctx_in {
>> > /** AMDGPU_CTX_OP_* */
>> > __u32 op;
>> > - /** For future use, no flags defined so far */
>> > + /** AMDGPU_CTX_FLAG_* */
>> > __u32 flags;
>> > __u32 ctx_id;
>> > __u32 _pad;
>> > --
>> > 2.7.4
>> >
>> > _______________________________________________
>> > amd-gfx mailing list
>> > amd-gfx at lists.freedesktop.org
>> > https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>
>
More information about the amd-gfx
mailing list