[PATCH 3/3] drm/amdgpu: add FENCE_TO_HANDLE ioctl that returns syncobj or sync_file

zhoucm1 david1.zhou at amd.com
Wed Sep 13 07:39:04 UTC 2017



On 2017年09月13日 15:09, Christian König wrote:
> syncfile is the older implementation, sync_obj is the replacement for it.

Are you sure sync_obj is a replacement for syncfile? Anyone can clarify it?

sync_obj is mainly for semaphore usage, we can see sync_obj docs from 
Dave in drm_syncobj.c:
"/**
  * DOC: Overview
  *
  * DRM synchronisation objects (syncobj) are a persistent objects,
  * that contain an optional fence. The fence can be updated with a new
  * fence, or be NULL.
  *
  * syncobj's can be export to fd's and back, these fd's are opaque and
  * have no other use case, except passing the syncobj between processes.
  *
  * Their primary use-case is to implement Vulkan fences and semaphores.
  *
  * syncobj have a kref reference count, but also have an optional file.
  * The file is only created once the syncobj is exported.
  * The file takes a reference on the kref.
  */
"

>
> I don't think we should implement syncfile in the CS any more, 
> sync_obj is already done and can be converted to/from syncfiles.
>
> Mareks IOCTL should cover the case when we need to create a syncfile 
> or syncobj from a fence sequence number.

I know his convertion can do that things, but returning syncfile fd 
directly is a really simple method.

Regards,
David Zhou
>
> Regards,
> Christian.
>
> Am 13.09.2017 um 09:03 schrieb zhoucm1:
>> I really very surprise that you prefer to expand sync_obj to get 
>> syncfile fd!!!
>>
>> Why not directly generate syncfile fd and use it? Which one is simple 
>> is so obvious.
>>
>> Regards,
>> David Zhou
>> On 2017年09月13日 14:57, Christian König wrote:
>>> Hi guys,
>>>
>>> Mareks IOCTL proposal looks really good to me.
>>>
>>> Please note that we already have sync_obj support for the CS IOCTL 
>>> in the 4.13 branch and this work is based on top of that.
>>>
>>>> UMD don't need to construct ip_type/ip_instance/ctx_id/ring
>>> Well the UMD does want to construct ip_type/ip_instance/ctx_id/ring 
>>> and use for the simple reason that it allows more flexibility than 
>>> sync_obj/sync_file.
>>>
>>> Thinking more about this I'm pretty sure we want to do something 
>>> similar for VM map/unmap operations as well.
>>>
>>> Regards,
>>> Christian.
>>>
>>> Am 13.09.2017 um 05:03 schrieb zhoucm1:
>>>> Hi Marek,
>>>>
>>>> You're doing same things with me, see my "introduce syncfile as 
>>>> fence reuturn" patch set, which makes things more simple, we just 
>>>> need to directly return syncfile fd to UMD when CS, then the fence 
>>>> UMD get will be always syncfile fd, UMD don't need to construct 
>>>> ip_type/ip_instance/ctx_id/ring any more, which also can pass to 
>>>> dependency and syncobj as well.
>>>>
>>>> Regards,
>>>> David Zhou
>>>> On 2017年09月13日 04:42, Marek Olšák wrote:
>>>>> From: Marek Olšák <marek.olsak at amd.com>
>>>>>
>>>>> for being able to convert an amdgpu fence into one of the handles.
>>>>> Mesa will use this.
>>>>>
>>>>> Signed-off-by: Marek Olšák <marek.olsak at amd.com>
>>>>> ---
>>>>>   drivers/gpu/drm/amd/amdgpu/amdgpu.h     |  2 ++
>>>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c  | 61 
>>>>> +++++++++++++++++++++++++++++++++
>>>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c |  3 +-
>>>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c |  1 +
>>>>>   include/uapi/drm/amdgpu_drm.h           | 16 +++++++++
>>>>>   5 files changed, 82 insertions(+), 1 deletion(-)
>>>>>
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h 
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> index b5c8b90..c15fa93 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h
>>>>> @@ -1308,6 +1308,8 @@ int amdgpu_gem_va_ioctl(struct drm_device 
>>>>> *dev, void *data,
>>>>>   int amdgpu_gem_op_ioctl(struct drm_device *dev, void *data,
>>>>>               struct drm_file *filp);
>>>>>   int amdgpu_cs_ioctl(struct drm_device *dev, void *data, struct 
>>>>> drm_file *filp);
>>>>> +int amdgpu_cs_fence_to_handle_ioctl(struct drm_device *dev, void 
>>>>> *data,
>>>>> +                    struct drm_file *filp);
>>>>>   int amdgpu_cs_wait_ioctl(struct drm_device *dev, void *data, 
>>>>> struct drm_file *filp);
>>>>>   int amdgpu_cs_wait_fences_ioctl(struct drm_device *dev, void *data,
>>>>>                   struct drm_file *filp);
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
>>>>> index 7cb8a59..6dd719c 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
>>>>> @@ -25,6 +25,7 @@
>>>>>    *    Jerome Glisse <glisse at freedesktop.org>
>>>>>    */
>>>>>   #include <linux/pagemap.h>
>>>>> +#include <linux/sync_file.h>
>>>>>   #include <drm/drmP.h>
>>>>>   #include <drm/amdgpu_drm.h>
>>>>>   #include <drm/drm_syncobj.h>
>>>>> @@ -1311,6 +1312,66 @@ static struct dma_fence 
>>>>> *amdgpu_cs_get_fence(struct amdgpu_device *adev,
>>>>>       return fence;
>>>>>   }
>>>>>   +int amdgpu_cs_fence_to_handle_ioctl(struct drm_device *dev, 
>>>>> void *data,
>>>>> +                    struct drm_file *filp)
>>>>> +{
>>>>> +    struct amdgpu_device *adev = dev->dev_private;
>>>>> +    struct amdgpu_fpriv *fpriv = filp->driver_priv;
>>>>> +    union drm_amdgpu_fence_to_handle *info = data;
>>>>> +    struct dma_fence *fence;
>>>>> +    struct drm_syncobj *syncobj;
>>>>> +    struct sync_file *sync_file;
>>>>> +    int fd, r;
>>>>> +
>>>>> +    if (amdgpu_kms_vram_lost(adev, fpriv))
>>>>> +        return -ENODEV;
>>>>> +
>>>>> +    fence = amdgpu_cs_get_fence(adev, filp, &info->in.fence);
>>>>> +    if (IS_ERR(fence))
>>>>> +        return PTR_ERR(fence);
>>>>> +
>>>>> +    switch (info->in.what) {
>>>>> +    case AMDGPU_FENCE_TO_HANDLE_GET_SYNCOBJ:
>>>>> +        r = drm_syncobj_create(&syncobj, 0, fence);
>>>>> +        dma_fence_put(fence);
>>>>> +        if (r)
>>>>> +            return r;
>>>>> +        r = drm_syncobj_get_handle(filp, syncobj, 
>>>>> &info->out.handle);
>>>>> +        drm_syncobj_put(syncobj);
>>>>> +        return r;
>>>>> +
>>>>> +    case AMDGPU_FENCE_TO_HANDLE_GET_SYNCOBJ_FD:
>>>>> +        r = drm_syncobj_create(&syncobj, 0, fence);
>>>>> +        dma_fence_put(fence);
>>>>> +        if (r)
>>>>> +            return r;
>>>>> +        r = drm_syncobj_get_fd(syncobj, (int*)&info->out.handle);
>>>>> +        drm_syncobj_put(syncobj);
>>>>> +        return r;
>>>>> +
>>>>> +    case AMDGPU_FENCE_TO_HANDLE_GET_SYNC_FILE_FD:
>>>>> +        fd = get_unused_fd_flags(O_CLOEXEC);
>>>>> +        if (fd < 0) {
>>>>> +            dma_fence_put(fence);
>>>>> +            return fd;
>>>>> +        }
>>>>> +
>>>>> +        sync_file = sync_file_create(fence);
>>>>> +        dma_fence_put(fence);
>>>>> +        if (!sync_file) {
>>>>> +            put_unused_fd(fd);
>>>>> +            return -ENOMEM;
>>>>> +        }
>>>>> +
>>>>> +        fd_install(fd, sync_file->file);
>>>>> +        info->out.handle = fd;
>>>>> +        return 0;
>>>>> +
>>>>> +    default:
>>>>> +        return -EINVAL;
>>>>> +    }
>>>>> +}
>>>>> +
>>>>>   /**
>>>>>    * amdgpu_cs_wait_all_fence - wait on all fences to signal
>>>>>    *
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c 
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c
>>>>> index d01aca6..1e38411 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c
>>>>> @@ -70,9 +70,10 @@
>>>>>    * - 3.18.0 - Export gpu always on cu bitmap
>>>>>    * - 3.19.0 - Add support for UVD MJPEG decode
>>>>>    * - 3.20.0 - Add support for local BOs
>>>>> + * - 3.21.0 - Add DRM_AMDGPU_FENCE_TO_HANDLE ioctl
>>>>>    */
>>>>>   #define KMS_DRIVER_MAJOR    3
>>>>> -#define KMS_DRIVER_MINOR    20
>>>>> +#define KMS_DRIVER_MINOR    21
>>>>>   #define KMS_DRIVER_PATCHLEVEL    0
>>>>>     int amdgpu_vram_limit = 0;
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
>>>>> index d31777b..b09d315 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
>>>>> @@ -1021,6 +1021,7 @@ const struct drm_ioctl_desc 
>>>>> amdgpu_ioctls_kms[] = {
>>>>>       DRM_IOCTL_DEF_DRV(AMDGPU_CTX, amdgpu_ctx_ioctl, 
>>>>> DRM_AUTH|DRM_RENDER_ALLOW),
>>>>>       DRM_IOCTL_DEF_DRV(AMDGPU_VM, amdgpu_vm_ioctl, 
>>>>> DRM_AUTH|DRM_RENDER_ALLOW),
>>>>>       DRM_IOCTL_DEF_DRV(AMDGPU_BO_LIST, amdgpu_bo_list_ioctl, 
>>>>> DRM_AUTH|DRM_RENDER_ALLOW),
>>>>> +    DRM_IOCTL_DEF_DRV(AMDGPU_FENCE_TO_HANDLE, 
>>>>> amdgpu_cs_fence_to_handle_ioctl, DRM_AUTH|DRM_RENDER_ALLOW),
>>>>>       /* KMS */
>>>>>       DRM_IOCTL_DEF_DRV(AMDGPU_GEM_MMAP, amdgpu_gem_mmap_ioctl, 
>>>>> DRM_AUTH|DRM_RENDER_ALLOW),
>>>>>       DRM_IOCTL_DEF_DRV(AMDGPU_GEM_WAIT_IDLE, 
>>>>> amdgpu_gem_wait_idle_ioctl, DRM_AUTH|DRM_RENDER_ALLOW),
>>>>> diff --git a/include/uapi/drm/amdgpu_drm.h 
>>>>> b/include/uapi/drm/amdgpu_drm.h
>>>>> index 9f5bd97..ec57cd0 100644
>>>>> --- a/include/uapi/drm/amdgpu_drm.h
>>>>> +++ b/include/uapi/drm/amdgpu_drm.h
>>>>> @@ -53,6 +53,7 @@ extern "C" {
>>>>>   #define DRM_AMDGPU_WAIT_FENCES        0x12
>>>>>   #define DRM_AMDGPU_VM            0x13
>>>>>   #define DRM_AMDGPU_FREESYNC            0x14
>>>>> +#define DRM_AMDGPU_FENCE_TO_HANDLE    0x15
>>>>>     #define DRM_IOCTL_AMDGPU_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE 
>>>>> + DRM_AMDGPU_GEM_CREATE, union drm_amdgpu_gem_create)
>>>>>   #define DRM_IOCTL_AMDGPU_GEM_MMAP DRM_IOWR(DRM_COMMAND_BASE + 
>>>>> DRM_AMDGPU_GEM_MMAP, union drm_amdgpu_gem_mmap)
>>>>> @@ -69,6 +70,7 @@ extern "C" {
>>>>>   #define DRM_IOCTL_AMDGPU_WAIT_FENCES DRM_IOWR(DRM_COMMAND_BASE + 
>>>>> DRM_AMDGPU_WAIT_FENCES, union drm_amdgpu_wait_fences)
>>>>>   #define DRM_IOCTL_AMDGPU_VM DRM_IOWR(DRM_COMMAND_BASE + 
>>>>> DRM_AMDGPU_VM, union drm_amdgpu_vm)
>>>>>   #define DRM_IOCTL_AMDGPU_FREESYNC DRM_IOWR(DRM_COMMAND_BASE + 
>>>>> DRM_AMDGPU_FREESYNC, struct drm_amdgpu_freesync)
>>>>> +#define DRM_IOCTL_AMDGPU_FENCE_TO_HANDLE 
>>>>> DRM_IOWR(DRM_COMMAND_BASE + DRM_AMDGPU_FENCE_TO_HANDLE, union 
>>>>> drm_amdgpu_fence_to_handle)
>>>>>     #define AMDGPU_GEM_DOMAIN_CPU        0x1
>>>>>   #define AMDGPU_GEM_DOMAIN_GTT        0x2
>>>>> @@ -517,6 +519,20 @@ struct drm_amdgpu_cs_chunk_sem {
>>>>>       __u32 handle;
>>>>>   };
>>>>>   +#define AMDGPU_FENCE_TO_HANDLE_GET_SYNCOBJ    0
>>>>> +#define AMDGPU_FENCE_TO_HANDLE_GET_SYNCOBJ_FD    1
>>>>> +#define AMDGPU_FENCE_TO_HANDLE_GET_SYNC_FILE_FD    2
>>>>> +
>>>>> +union drm_amdgpu_fence_to_handle {
>>>>> +    struct {
>>>>> +        struct drm_amdgpu_fence fence;
>>>>> +        __u32 what;
>>>>> +    } in;
>>>>> +    struct {
>>>>> +        __u32 handle;
>>>>> +    } out;
>>>>> +};
>>>>> +
>>>>>   struct drm_amdgpu_cs_chunk_data {
>>>>>       union {
>>>>>           struct drm_amdgpu_cs_chunk_ib        ib_data;
>>>>
>>>> _______________________________________________
>>>> amd-gfx mailing list
>>>> amd-gfx at lists.freedesktop.org
>>>> https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>>>
>>>
>>
>> _______________________________________________
>> amd-gfx mailing list
>> amd-gfx at lists.freedesktop.org
>> https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>
>



More information about the dri-devel mailing list