[RFC PATCH 16/18] drm/amdgpu: Implement SET_PRIORITY GEM op
Friedrich Vock
friedrich.vock at gmx.de
Thu Apr 25 07:06:08 UTC 2024
On 25.04.24 08:58, Christian König wrote:
> Am 25.04.24 um 08:46 schrieb Friedrich Vock:
>> On 25.04.24 08:32, Christian König wrote:
>>> Am 24.04.24 um 18:57 schrieb Friedrich Vock:
>>>> Used by userspace to adjust buffer priorities in response to
>>>> changes in
>>>> application demand and memory pressure.
>>>
>>> Yeah, that was discussed over and over again. One big design criteria
>>> is that we can't have global priorities from userspace!
>>>
>>> The background here is that this can trivially be abused.
>>>
>> I see your point when apps are allowed to prioritize themselves above
>> other apps, and I agree that should probably be disallowed at least for
>> unprivileged apps.
>>
>> Disallowing this is a pretty trivial change though, and I don't really
>> see the abuse potential in being able to downgrade your own priority?
>
> Yeah, I know what you mean and I'm also leaning towards that
> argumentation. But another good point is also that it doesn't actually
> help.
>
> For example when you have desktop apps fighting with a game, you
> probably don't want to use static priorities, but rather evict the
> apps which are inactive and keep the apps which are active in the
> background.
>
Sadly things are not as simple as "evict everything from app 1, keep
everything from app 2 active". The simplest failure case of this is
games that already oversubscribe VRAM on their own. Keeping the whole
app inside VRAM is literally impossible there, and it helps a lot to
know which buffers the app is most happy with evicting.
> In other words the priority just tells you which stuff from each app
> to evict first, but not which app to globally throw out.
>
Yeah, but per-buffer priority system could do both of these.
Regards,
Friedrich
> Regards,
> Christian.
>
>>
>> Regards,
>> Friedrich
>>
>>> What we can do is to have per process priorities, but that needs to be
>>> in the VM subsystem.
>>>
>>> That's also the reason why I personally think that the handling
>>> shouldn't be inside TTM at all.
>>>
>>> Regards,
>>> Christian.
>>>
>>>>
>>>> Signed-off-by: Friedrich Vock <friedrich.vock at gmx.de>
>>>> ---
>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c | 20 ++++++++++++++++++++
>>>> include/uapi/drm/amdgpu_drm.h | 1 +
>>>> 2 files changed, 21 insertions(+)
>>>>
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
>>>> index 5ca13e2e50f50..6107810a9c205 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
>>>> @@ -836,8 +836,10 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev,
>>>> void *data,
>>>> {
>>>> struct amdgpu_device *adev = drm_to_adev(dev);
>>>> struct drm_amdgpu_gem_op *args = data;
>>>> + struct ttm_resource_manager *man;
>>>> struct drm_gem_object *gobj;
>>>> struct amdgpu_vm_bo_base *base;
>>>> + struct ttm_operation_ctx ctx;
>>>> struct amdgpu_bo *robj;
>>>> int r;
>>>>
>>>> @@ -851,6 +853,9 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev,
>>>> void *data,
>>>> if (unlikely(r))
>>>> goto out;
>>>>
>>>> + memset(&ctx, 0, sizeof(ctx));
>>>> + ctx.interruptible = true;
>>>> +
>>>> switch (args->op) {
>>>> case AMDGPU_GEM_OP_GET_GEM_CREATE_INFO: {
>>>> struct drm_amdgpu_gem_create_in info;
>>>> @@ -898,6 +903,21 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev,
>>>> void *data,
>>>>
>>>> amdgpu_bo_unreserve(robj);
>>>> break;
>>>> + case AMDGPU_GEM_OP_SET_PRIORITY:
>>>> + if (args->value > AMDGPU_BO_PRIORITY_MAX_USER)
>>>> + args->value = AMDGPU_BO_PRIORITY_MAX_USER;
>>>> + ttm_bo_update_priority(&robj->tbo, args->value);
>>>> + if (robj->tbo.evicted_type != TTM_NUM_MEM_TYPES) {
>>>> + ttm_bo_try_unevict(&robj->tbo, &ctx);
>>>> + amdgpu_bo_unreserve(robj);
>>>> + } else {
>>>> + amdgpu_bo_unreserve(robj);
>>>> + man = ttm_manager_type(robj->tbo.bdev,
>>>> + robj->tbo.resource->mem_type);
>>>> + ttm_mem_unevict_evicted(robj->tbo.bdev, man,
>>>> + true);
>>>> + }
>>>> + break;
>>>> default:
>>>> amdgpu_bo_unreserve(robj);
>>>> r = -EINVAL;
>>>> diff --git a/include/uapi/drm/amdgpu_drm.h
>>>> b/include/uapi/drm/amdgpu_drm.h
>>>> index bdbe6b262a78d..53552dd489b9b 100644
>>>> --- a/include/uapi/drm/amdgpu_drm.h
>>>> +++ b/include/uapi/drm/amdgpu_drm.h
>>>> @@ -531,6 +531,7 @@ union drm_amdgpu_wait_fences {
>>>>
>>>> #define AMDGPU_GEM_OP_GET_GEM_CREATE_INFO 0
>>>> #define AMDGPU_GEM_OP_SET_PLACEMENT 1
>>>> +#define AMDGPU_GEM_OP_SET_PRIORITY 2
>>>>
>>>> /* Sets or returns a value associated with a buffer. */
>>>> struct drm_amdgpu_gem_op {
>>>> --
>>>> 2.44.0
>>>>
>>>
>
More information about the dri-devel
mailing list