[PATCH 1/3] [RFC]drm: add syncobj timeline support v4
zhoucm1
zhoucm1 at amd.com
Thu Sep 13 02:15:02 UTC 2018
On 2018年09月12日 19:05, Christian König wrote:
> Am 12.09.2018 um 12:20 schrieb zhoucm1:
>> [SNIP]
>>>> Drop the term semaphore here, better use syncobj.
>> This is from VK_KHR_timeline_semaphore extension describe, not my
>> invention, I just quote it. In kernel side, we call syncobj, in UMD,
>> they still call semaphore.
>
> Yeah, but we don't care about close source UMD names in the kernel and'
> the open source UMD calls it syncobj as well.
>
>>>>
>>>>> [SNIP]
>>>>> +static void drm_syncobj_find_signal_pt_for_wait_pt(struct
>>>>> drm_syncobj *syncobj,
>>>>> + struct drm_syncobj_wait_pt *wait_pt)
>>>>> +{
>>>>
>>>> That whole approach still looks horrible complicated to me.
>> It's already very close to what you said before.
>>
>>>>
>>>> Especially the separation of signal and wait pt is completely
>>>> unnecessary as far as I can see.
>>>> When a wait pt is requested we just need to search for the signal
>>>> point which it will trigger.
>> Yeah, I tried this, but when I implement cpu wait ioctl on specific
>> point, we need a advanced wait pt fence, otherwise, we could still
>> need old syncobj cb.
>
> Why? I mean you just need to call drm_syncobj_find_fence() and when
> that one returns NULL you use wait_event_*() to wait for a signal
> point >= your wait point to appear and try again.
e.g. when there are 3 syncobjs(A,B,C) to wait, all syncobjABC have no
fence yet, as you said, during drm_syncobj_find_fence(A) is working on
wait_event, syncobjB and syncobjC could already be signaled, then we
don't know which one is first signaled, which is need when wait ioctl
returns.
Back to my implementation, it already fixes all your concerns before,
and can be able to easily used in wait_ioctl. When you feel that is
complicated, I guess that is because we merged all logic to that and
much clean up in one patch. In fact, it already is very simple,
timeline_init/fini, create signal/wait_pt, find signal_pt for wait_pt,
garbage collection, just them.
Thanks,
David Zhou
>
> Regards,
> Christian.
>
>>
>>
>> Thanks,
>> David Zhou
>>>>
>>>> Regards,
>>>> Christian.
>>>>
>>>>> + struct drm_syncobj_timeline *timeline =
>>>>> &syncobj->syncobj_timeline;
>>>>> + struct drm_syncobj_signal_pt *signal_pt;
>>>>> + int ret;
>>>>> +
>>>>> + if (wait_pt->signal_pt_fence) {
>>>>> + return;
>>>>> + } else if ((syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) &&
>>>>> + (wait_pt->value <= timeline->timeline)) {
>>>>> + dma_fence_signal(&wait_pt->base.base);
>>>>> + rb_erase(&wait_pt->node,
>>>>> + &timeline->wait_pt_tree);
>>>>> + RB_CLEAR_NODE(&wait_pt->node);
>>>>> + dma_fence_put(&wait_pt->base.base);
>>>>> + return;
>>>>> + }
>>>>> +
>>>>> + list_for_each_entry(signal_pt, &timeline->signal_pt_list,
>>>>> list) {
>>>>> + if (wait_pt->value < signal_pt->value)
>>>>> + continue;
>>>>> + if ((syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) &&
>>>>> + (wait_pt->value != signal_pt->value))
>>>>> + continue;
>>>>> + wait_pt->signal_pt_fence =
>>>>> dma_fence_get(&signal_pt->base->base);
>>>>> + ret = dma_fence_add_callback(wait_pt->signal_pt_fence,
>>>>> + &wait_pt->wait_cb,
>>>>> + wait_pt_func);
>>>>> + if (ret == -ENOENT) {
>>>>> + dma_fence_signal(&wait_pt->base.base);
>>>>> + dma_fence_put(wait_pt->signal_pt_fence);
>>>>> + wait_pt->signal_pt_fence = NULL;
>>>>> + rb_erase(&wait_pt->node,
>>>>> + &timeline->wait_pt_tree);
>>>>> + RB_CLEAR_NODE(&wait_pt->node);
>>>>> + dma_fence_put(&wait_pt->base.base);
>>>>> + } else if (ret < 0) {
>>>>> + dma_fence_put(wait_pt->signal_pt_fence);
>>>>> + DRM_ERROR("add callback error!");
>>>>> + } else {
>>>>> + /* after adding callback, remove from rb tree */
>>>>> + rb_erase(&wait_pt->node,
>>>>> + &timeline->wait_pt_tree);
>>>>> + RB_CLEAR_NODE(&wait_pt->node);
>>>>> + }
>>>>> + return;
>>>>> + }
>>>>> + /* signaled pt was released */
>>>>> + if (!wait_pt->signal_pt_fence && (wait_pt->value <=
>>>>> + timeline->signal_point)) {
>>>>> + dma_fence_signal(&wait_pt->base.base);
>>>>> + rb_erase(&wait_pt->node,
>>>>> + &timeline->wait_pt_tree);
>>>>> + RB_CLEAR_NODE(&wait_pt->node);
>>>>> + dma_fence_put(&wait_pt->base.base);
>>>>> + }
>>>>> }
>>>>> -void drm_syncobj_add_callback(struct drm_syncobj *syncobj,
>>>>> - struct drm_syncobj_cb *cb,
>>>>> - drm_syncobj_func_t func)
>>>>> +static int drm_syncobj_timeline_create_signal_pt(struct
>>>>> drm_syncobj *syncobj,
>>>>> + struct dma_fence *fence,
>>>>> + u64 point)
>>>>> {
>>>>> + struct drm_syncobj_signal_pt *signal_pt =
>>>>> + kzalloc(sizeof(struct drm_syncobj_signal_pt), GFP_KERNEL);
>>>>> + struct drm_syncobj_signal_pt *tail_pt;
>>>>> + struct dma_fence **fences;
>>>>> + struct rb_node *node;
>>>>> + struct drm_syncobj_wait_pt *tail_wait_pt = NULL;
>>>>> + int num_fences = 0;
>>>>> + int ret = 0, i;
>>>>> +
>>>>> + if (!signal_pt)
>>>>> + return -ENOMEM;
>>>>> + if (syncobj->syncobj_timeline.signal_point >= point) {
>>>>> + DRM_WARN("A later signal is ready!");
>>>>> + goto out;
>>>>> + }
>>>>> + if (!fence)
>>>>> + goto out;
>>>>> +
>>>>> + fences = kmalloc_array(sizeof(void *), 2, GFP_KERNEL);
>>>>> + if (!fences)
>>>>> + goto out;
>>>>> + fences[num_fences++] = dma_fence_get(fence);
>>>>> + /* timeline syncobj must take this dependency */
>>>>> + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>>>>> + spin_lock(&syncobj->lock);
>>>>> + if
>>>>> (!list_empty(&syncobj->syncobj_timeline.signal_pt_list)) {
>>>>> + tail_pt =
>>>>> list_last_entry(&syncobj->syncobj_timeline.signal_pt_list,
>>>>> + struct drm_syncobj_signal_pt, list);
>>>>> + fences[num_fences++] =
>>>>> dma_fence_get(&tail_pt->base->base);
>>>>> + }
>>>>> + spin_unlock(&syncobj->lock);
>>>>> + }
>>>>> + signal_pt->base = dma_fence_array_create(num_fences, fences,
>>>>> + syncobj->syncobj_timeline.timeline_context,
>>>>> + point, false);
>>>>> + if (!signal_pt->base)
>>>>> + goto fail;
>>>>> +
>>>>> + signal_pt->value = point;
>>>>> spin_lock(&syncobj->lock);
>>>>> - drm_syncobj_add_callback_locked(syncobj, cb, func);
>>>>> + syncobj->syncobj_timeline.signal_point = point;
>>>>> + INIT_LIST_HEAD(&signal_pt->list);
>>>>> + list_add_tail(&signal_pt->list,
>>>>> &syncobj->syncobj_timeline.signal_pt_list);
>>>>> +
>>>>> + /* check if there is advanced wait */
>>>>> + node = rb_last(&syncobj->syncobj_timeline.wait_pt_tree);
>>>>> + if (node)
>>>>> + tail_wait_pt = rb_entry(node, struct drm_syncobj_wait_pt,
>>>>> node);
>>>>> + if (tail_wait_pt && !tail_wait_pt->signal_pt_fence) {
>>>>> + for (node =
>>>>> rb_first(&syncobj->syncobj_timeline.wait_pt_tree);
>>>>> + node != NULL; node = rb_next(node)) {
>>>>> + struct drm_syncobj_wait_pt *wait_pt =
>>>>> + rb_entry(node, struct drm_syncobj_wait_pt,
>>>>> + node);
>>>>> +
>>>>> + drm_syncobj_find_signal_pt_for_wait_pt(syncobj,
>>>>> + wait_pt);
>>>>> + }
>>>>> + }
>>>>> +
>>>>> spin_unlock(&syncobj->lock);
>>>>> + wake_up_all(&syncobj->syncobj_timeline.wq);
>>>>> +
>>>>> + return 0;
>>>>> +fail:
>>>>> + for (i = 0; i < num_fences; i++)
>>>>> + dma_fence_put(fences[i]);
>>>>> + kfree(fences);
>>>>> +out:
>>>>> + kfree(signal_pt);
>>>>> + return ret;
>>>>> }
>>>>> -void drm_syncobj_remove_callback(struct drm_syncobj *syncobj,
>>>>> - struct drm_syncobj_cb *cb)
>>>>> +static void drm_syncobj_timeline_garbage_collection(struct
>>>>> drm_syncobj *syncobj)
>>>>> {
>>>>> + struct drm_syncobj_timeline *timeline =
>>>>> &syncobj->syncobj_timeline;
>>>>> + struct rb_node *node;
>>>>> + struct drm_syncobj_wait_pt *wait_pt;
>>>>> + struct drm_syncobj_signal_pt *signal_pt, *tmp;
>>>>> +
>>>>> spin_lock(&syncobj->lock);
>>>>> - list_del_init(&cb->node);
>>>>> + node = rb_first(&timeline->wait_pt_tree);
>>>>> + while (node) {
>>>>> + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>>>>> + if (!dma_fence_is_signaled(&wait_pt->base.base)) {
>>>>> + node = rb_next(node);
>>>>> + continue;
>>>>> + }
>>>>> + rb_erase(&wait_pt->node,
>>>>> + &timeline->wait_pt_tree);
>>>>> + RB_CLEAR_NODE(&wait_pt->node);
>>>>> + /* kfree(wait_pt) is excuted by fence put */
>>>>> + dma_fence_put(&wait_pt->base.base);
>>>>> + node = rb_first(&timeline->wait_pt_tree);
>>>>> + }
>>>>> + list_for_each_entry_safe(signal_pt, tmp,
>>>>> + &timeline->signal_pt_list, list) {
>>>>> + if (dma_fence_is_signaled(&signal_pt->base->base)) {
>>>>> + timeline->timeline = signal_pt->value;
>>>>> + list_del(&signal_pt->list);
>>>>> + dma_fence_put(&signal_pt->base->base);
>>>>> + kfree(signal_pt);
>>>>> + } else {
>>>>> + /*signal_pt is in order in list, from small to big, so
>>>>> + * the later must not be signal either */
>>>>> + break;
>>>>> + }
>>>>> + }
>>>>> +
>>>>> spin_unlock(&syncobj->lock);
>>>>> }
>>>>> -
>>>>> /**
>>>>> * drm_syncobj_replace_fence - replace fence in a sync object.
>>>>> * @syncobj: Sync object to replace fence in
>>>>> @@ -176,28 +368,29 @@ void drm_syncobj_replace_fence(struct
>>>>> drm_syncobj *syncobj,
>>>>> u64 point,
>>>>> struct dma_fence *fence)
>>>>> {
>>>>> - struct dma_fence *old_fence;
>>>>> - struct drm_syncobj_cb *cur, *tmp;
>>>>> -
>>>>> - if (fence)
>>>>> - dma_fence_get(fence);
>>>>> -
>>>>> - spin_lock(&syncobj->lock);
>>>>> -
>>>>> - old_fence = rcu_dereference_protected(syncobj->fence,
>>>>> - lockdep_is_held(&syncobj->lock));
>>>>> - rcu_assign_pointer(syncobj->fence, fence);
>>>>> -
>>>>> - if (fence != old_fence) {
>>>>> - list_for_each_entry_safe(cur, tmp, &syncobj->cb_list,
>>>>> node) {
>>>>> - list_del_init(&cur->node);
>>>>> - cur->func(syncobj, cur);
>>>>> + drm_syncobj_timeline_garbage_collection(syncobj);
>>>>> + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>>>>> + if (fence)
>>>>> + drm_syncobj_timeline_create_signal_pt(syncobj, fence,
>>>>> + point);
>>>>> + return;
>>>>> + } else if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) {
>>>>> + u64 pt_value;
>>>>> +
>>>>> + if (!fence) {
>>>>> + drm_syncobj_timeline_fini(syncobj,
>>>>> + &syncobj->syncobj_timeline);
>>>>> + drm_syncobj_timeline_init(syncobj,
>>>>> + &syncobj->syncobj_timeline);
>>>>> + return;
>>>>> }
>>>>> + pt_value = syncobj->syncobj_timeline.signal_point +
>>>>> + DRM_SYNCOBJ_NORMAL_POINT;
>>>>> + drm_syncobj_timeline_create_signal_pt(syncobj, fence,
>>>>> pt_value);
>>>>> + return;
>>>>> + } else {
>>>>> + DRM_ERROR("the syncobj type isn't support\n");
>>>>> }
>>>>> -
>>>>> - spin_unlock(&syncobj->lock);
>>>>> -
>>>>> - dma_fence_put(old_fence);
>>>>> }
>>>>> EXPORT_SYMBOL(drm_syncobj_replace_fence);
>>>>> @@ -220,6 +413,120 @@ static int
>>>>> drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj)
>>>>> return 0;
>>>>> }
>>>>> +static struct drm_syncobj_wait_pt *
>>>>> +drm_syncobj_timeline_lookup_wait_pt_fence(struct drm_syncobj
>>>>> *syncobj,
>>>>> + u64 point,
>>>>> + struct dma_fence **fence)
>>>>> +{
>>>>> + struct rb_node *node =
>>>>> syncobj->syncobj_timeline.wait_pt_tree.rb_node;
>>>>> + struct drm_syncobj_wait_pt *wait_pt = NULL;
>>>>> +
>>>>> + spin_lock(&syncobj->lock);
>>>>> + while(node) {
>>>>> + int result;
>>>>> +
>>>>> + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>>>>> + result = point - wait_pt->value;
>>>>> + if (result < 0) {
>>>>> + node = node->rb_left;
>>>>> + } else if (result > 0) {
>>>>> + node = node->rb_right;
>>>>> + } else {
>>>>> + if (fence)
>>>>> + *fence = dma_fence_get(&wait_pt->base.base);
>>>>> + break;
>>>>> + }
>>>>> + }
>>>>> + spin_unlock(&syncobj->lock);
>>>>> +
>>>>> + return wait_pt;
>>>>> +}
>>>>> +
>>>>> +static struct drm_syncobj_wait_pt *
>>>>> +drm_syncobj_timeline_create_wait_pt_and_return_fence(struct
>>>>> drm_syncobj *syncobj,
>>>>> + u64 point,
>>>>> + struct dma_fence **fence)
>>>>> +{
>>>>> + struct drm_syncobj_wait_pt *wait_pt = kzalloc(sizeof(*wait_pt),
>>>>> + GFP_KERNEL);
>>>>> + struct drm_syncobj_stub_fence *base;
>>>>> + struct rb_node **new =
>>>>> &(syncobj->syncobj_timeline.wait_pt_tree.rb_node), *parent = NULL;
>>>>> +
>>>>> + if (!wait_pt)
>>>>> + return NULL;
>>>>> + base = &wait_pt->base;
>>>>> + spin_lock_init(&base->lock);
>>>>> + dma_fence_init(&base->base,
>>>>> + &drm_syncobj_stub_fence_ops,
>>>>> + &base->lock,
>>>>> + syncobj->syncobj_timeline.timeline_context, point);
>>>>> + wait_pt->value = point;
>>>>> + wait_pt->signal_pt_fence = NULL;
>>>>> +
>>>>> + /* wait pt must be in an order, so that we can easily lookup
>>>>> and signal
>>>>> + * it */
>>>>> + spin_lock(&syncobj->lock);
>>>>> + while(*new) {
>>>>> + struct drm_syncobj_wait_pt *this =
>>>>> + rb_entry(*new, struct drm_syncobj_wait_pt, node);
>>>>> + int result = wait_pt->value - this->value;
>>>>> +
>>>>> + parent = *new;
>>>>> + if (result < 0)
>>>>> + new = &((*new)->rb_left);
>>>>> + else if (result > 0)
>>>>> + new = &((*new)->rb_right);
>>>>> + else
>>>>> + goto exist;
>>>>> + }
>>>>> + if (fence)
>>>>> + *fence = dma_fence_get(&wait_pt->base.base);
>>>>> + rb_link_node(&wait_pt->node, parent, new);
>>>>> + rb_insert_color(&wait_pt->node,
>>>>> &syncobj->syncobj_timeline.wait_pt_tree);
>>>>> + spin_unlock(&syncobj->lock);
>>>>> + return wait_pt;
>>>>> +exist:
>>>>> + spin_unlock(&syncobj->lock);
>>>>> + dma_fence_put(&wait_pt->base.base);
>>>>> + wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj,
>>>>> point,
>>>>> + fence);
>>>>> + return wait_pt;
>>>>> +}
>>>>> +
>>>>> +static struct dma_fence *
>>>>> +drm_syncobj_timeline_point_get(struct drm_syncobj *syncobj, u64
>>>>> point, u64 flags)
>>>>> +{
>>>>> + struct drm_syncobj_wait_pt *wait_pt;
>>>>> + struct dma_fence *fence = NULL;
>>>>> +
>>>>> + /* check if the wait pt exists */
>>>>> + wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj,
>>>>> point,
>>>>> + &fence);
>>>>> + if (!fence) {
>>>>> + /* This is a new wait pt, so create it */
>>>>> + wait_pt =
>>>>> drm_syncobj_timeline_create_wait_pt_and_return_fence(syncobj, point,
>>>>> + &fence);
>>>>> + if (!fence)
>>>>> + return NULL;
>>>>> + }
>>>>> + if (wait_pt) {
>>>>> + int ret = 0;
>>>>> +
>>>>> + if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>>>>> + ret =
>>>>> wait_event_interruptible_timeout(syncobj->syncobj_timeline.wq,
>>>>> + wait_pt->value <=
>>>>> syncobj->syncobj_timeline.signal_point,
>>>>> + msecs_to_jiffies(10000)); /* wait 10s */
>>>>> + if (ret <= 0)
>>>>> + return NULL;
>>>>> + }
>>>>> + spin_lock(&syncobj->lock);
>>>>> + drm_syncobj_find_signal_pt_for_wait_pt(syncobj, wait_pt);
>>>>> + spin_unlock(&syncobj->lock);
>>>>> + return fence;
>>>>> + }
>>>>> + return NULL;
>>>>> +}
>>>>> +
>>>>> /**
>>>>> * drm_syncobj_find_fence - lookup and reference the fence in a
>>>>> sync object
>>>>> * @file_private: drm file private pointer
>>>>> @@ -234,20 +541,46 @@ static int
>>>>> drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj)
>>>>> * contains a reference to the fence, which must be released by
>>>>> calling
>>>>> * dma_fence_put().
>>>>> */
>>>>> -int drm_syncobj_find_fence(struct drm_file *file_private,
>>>>> - u32 handle, u64 point,
>>>>> - struct dma_fence **fence)
>>>>> +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point,
>>>>> + u64 flags, struct dma_fence **fence)
>>>>> {
>>>>> - struct drm_syncobj *syncobj = drm_syncobj_find(file_private,
>>>>> handle);
>>>>> int ret = 0;
>>>>> if (!syncobj)
>>>>> return -ENOENT;
>>>>> - *fence = drm_syncobj_fence_get(syncobj);
>>>>> + drm_syncobj_timeline_garbage_collection(syncobj);
>>>>> + if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) {
>>>>> + /*NORMAL syncobj always wait on last pt */
>>>>> + u64 tail_pt_value = syncobj->syncobj_timeline.signal_point;
>>>>> +
>>>>> + if (tail_pt_value == 0)
>>>>> + tail_pt_value += DRM_SYNCOBJ_NORMAL_POINT;
>>>>> + /* NORMAL syncobj doesn't care point value */
>>>>> + WARN_ON(point != 0);
>>>>> + *fence = drm_syncobj_timeline_point_get(syncobj,
>>>>> tail_pt_value,
>>>>> + flags);
>>>>> + } else if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>>>>> + *fence = drm_syncobj_timeline_point_get(syncobj, point,
>>>>> + flags);
>>>>> + } else {
>>>>> + DRM_ERROR("Don't support this type syncobj\n");
>>>>> + *fence = NULL;
>>>>> + }
>>>>> if (!*fence) {
>>>>> ret = -EINVAL;
>>>>> }
>>>>> + return ret;
>>>>> +}
>>>>> +EXPORT_SYMBOL(drm_syncobj_search_fence);
>>>>> +int drm_syncobj_find_fence(struct drm_file *file_private,
>>>>> + u32 handle, u64 point,
>>>>> + struct dma_fence **fence) {
>>>>> + struct drm_syncobj *syncobj = drm_syncobj_find(file_private,
>>>>> handle);
>>>>> +
>>>>> + int ret = drm_syncobj_search_fence(syncobj, point,
>>>>> + DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT,
>>>>> + fence);
>>>>> drm_syncobj_put(syncobj);
>>>>> return ret;
>>>>> }
>>>>> @@ -264,7 +597,7 @@ void drm_syncobj_free(struct kref *kref)
>>>>> struct drm_syncobj *syncobj = container_of(kref,
>>>>> struct drm_syncobj,
>>>>> refcount);
>>>>> - drm_syncobj_replace_fence(syncobj, 0, NULL);
>>>>> + drm_syncobj_timeline_fini(syncobj, &syncobj->syncobj_timeline);
>>>>> kfree(syncobj);
>>>>> }
>>>>> EXPORT_SYMBOL(drm_syncobj_free);
>>>>> @@ -292,8 +625,12 @@ int drm_syncobj_create(struct drm_syncobj
>>>>> **out_syncobj, uint32_t flags,
>>>>> return -ENOMEM;
>>>>> kref_init(&syncobj->refcount);
>>>>> - INIT_LIST_HEAD(&syncobj->cb_list);
>>>>> spin_lock_init(&syncobj->lock);
>>>>> + if (flags & DRM_SYNCOBJ_CREATE_TYPE_TIMELINE)
>>>>> + syncobj->type = DRM_SYNCOBJ_TYPE_TIMELINE;
>>>>> + else
>>>>> + syncobj->type = DRM_SYNCOBJ_TYPE_NORMAL;
>>>>> + drm_syncobj_timeline_init(syncobj, &syncobj->syncobj_timeline);
>>>>> if (flags & DRM_SYNCOBJ_CREATE_SIGNALED) {
>>>>> ret = drm_syncobj_assign_null_handle(syncobj);
>>>>> @@ -651,7 +988,6 @@ struct syncobj_wait_entry {
>>>>> struct task_struct *task;
>>>>> struct dma_fence *fence;
>>>>> struct dma_fence_cb fence_cb;
>>>>> - struct drm_syncobj_cb syncobj_cb;
>>>>> };
>>>>> static void syncobj_wait_fence_func(struct dma_fence *fence,
>>>>> @@ -663,18 +999,6 @@ static void syncobj_wait_fence_func(struct
>>>>> dma_fence *fence,
>>>>> wake_up_process(wait->task);
>>>>> }
>>>>> -static void syncobj_wait_syncobj_func(struct drm_syncobj *syncobj,
>>>>> - struct drm_syncobj_cb *cb)
>>>>> -{
>>>>> - struct syncobj_wait_entry *wait =
>>>>> - container_of(cb, struct syncobj_wait_entry, syncobj_cb);
>>>>> -
>>>>> - /* This happens inside the syncobj lock */
>>>>> - wait->fence =
>>>>> dma_fence_get(rcu_dereference_protected(syncobj->fence,
>>>>> - lockdep_is_held(&syncobj->lock)));
>>>>> - wake_up_process(wait->task);
>>>>> -}
>>>>> -
>>>>> static signed long drm_syncobj_array_wait_timeout(struct
>>>>> drm_syncobj **syncobjs,
>>>>> uint32_t count,
>>>>> uint32_t flags,
>>>>> @@ -698,14 +1022,11 @@ static signed long
>>>>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>>>> signaled_count = 0;
>>>>> for (i = 0; i < count; ++i) {
>>>>> entries[i].task = current;
>>>>> - entries[i].fence = drm_syncobj_fence_get(syncobjs[i]);
>>>>> + ret = drm_syncobj_search_fence(syncobjs[i], 0, 0,
>>>>> + &entries[i].fence);
>>>>> if (!entries[i].fence) {
>>>>> - if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>>>>> - continue;
>>>>> - } else {
>>>>> - ret = -EINVAL;
>>>>> - goto cleanup_entries;
>>>>> - }
>>>>> + ret = -EINVAL;
>>>>> + goto cleanup_entries;
>>>>> }
>>>>> if (dma_fence_is_signaled(entries[i].fence)) {
>>>>> @@ -733,15 +1054,6 @@ static signed long
>>>>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>>>> * fallthough and try a 0 timeout wait!
>>>>> */
>>>>> - if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>>>>> - for (i = 0; i < count; ++i) {
>>>>> - drm_syncobj_fence_get_or_add_callback(syncobjs[i],
>>>>> - &entries[i].fence,
>>>>> - &entries[i].syncobj_cb,
>>>>> - syncobj_wait_syncobj_func);
>>>>> - }
>>>>> - }
>>>>> -
>>>>> do {
>>>>> set_current_state(TASK_INTERRUPTIBLE);
>>>>> @@ -789,13 +1101,10 @@ static signed long
>>>>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>>>> cleanup_entries:
>>>>> for (i = 0; i < count; ++i) {
>>>>> - if (entries[i].syncobj_cb.func)
>>>>> - drm_syncobj_remove_callback(syncobjs[i],
>>>>> - &entries[i].syncobj_cb);
>>>>> + dma_fence_put(entries[i].fence);
>>>>> if (entries[i].fence_cb.func)
>>>>> dma_fence_remove_callback(entries[i].fence,
>>>>> &entries[i].fence_cb);
>>>>> - dma_fence_put(entries[i].fence);
>>>>> }
>>>>> kfree(entries);
>>>>> @@ -970,12 +1279,21 @@ drm_syncobj_reset_ioctl(struct drm_device
>>>>> *dev, void *data,
>>>>> if (ret < 0)
>>>>> return ret;
>>>>> - for (i = 0; i < args->count_handles; i++)
>>>>> - drm_syncobj_replace_fence(syncobjs[i], 0, NULL);
>>>>> -
>>>>> + for (i = 0; i < args->count_handles; i++) {
>>>>> + if (syncobjs[i]->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>>>>> + DRM_ERROR("timeline syncobj cannot reset!\n");
>>>>> + ret = -EINVAL;
>>>>> + goto out;
>>>>> + }
>>>>> + drm_syncobj_timeline_fini(syncobjs[i],
>>>>> + &syncobjs[i]->syncobj_timeline);
>>>>> + drm_syncobj_timeline_init(syncobjs[i],
>>>>> + &syncobjs[i]->syncobj_timeline);
>>>>> + }
>>>>> +out:
>>>>> drm_syncobj_array_free(syncobjs, args->count_handles);
>>>>> - return 0;
>>>>> + return ret;
>>>>> }
>>>>> int
>>>>> diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>>>>> b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>>>>> index 0a8d2d64f380..579e91a5858b 100644
>>>>> --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>>>>> +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>>>>> @@ -2137,7 +2137,9 @@ await_fence_array(struct i915_execbuffer *eb,
>>>>> if (!(flags & I915_EXEC_FENCE_WAIT))
>>>>> continue;
>>>>> - fence = drm_syncobj_fence_get(syncobj);
>>>>> + drm_syncobj_search_fence(syncobj, 0,
>>>>> + DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT,
>>>>> + &fence);
>>>>> if (!fence)
>>>>> return -EINVAL;
>>>>> diff --git a/include/drm/drm_syncobj.h b/include/drm/drm_syncobj.h
>>>>> index 425432b85a87..657c97dc25ec 100644
>>>>> --- a/include/drm/drm_syncobj.h
>>>>> +++ b/include/drm/drm_syncobj.h
>>>>> @@ -30,6 +30,25 @@
>>>>> struct drm_syncobj_cb;
>>>>> +enum drm_syncobj_type {
>>>>> + DRM_SYNCOBJ_TYPE_NORMAL,
>>>>> + DRM_SYNCOBJ_TYPE_TIMELINE
>>>>> +};
>>>>> +
>>>>> +struct drm_syncobj_timeline {
>>>>> + wait_queue_head_t wq;
>>>>> + u64 timeline_context;
>>>>> + /**
>>>>> + * @timeline: syncobj timeline
>>>>> + */
>>>>> + u64 timeline;
>>>>> + u64 signal_point;
>>>>> +
>>>>> +
>>>>> + struct rb_root wait_pt_tree;
>>>>> + struct list_head signal_pt_list;
>>>>> +};
>>>>> +
>>>>> /**
>>>>> * struct drm_syncobj - sync object.
>>>>> *
>>>>> @@ -41,19 +60,16 @@ struct drm_syncobj {
>>>>> */
>>>>> struct kref refcount;
>>>>> /**
>>>>> - * @fence:
>>>>> - * NULL or a pointer to the fence bound to this object.
>>>>> - *
>>>>> - * This field should not be used directly. Use
>>>>> drm_syncobj_fence_get()
>>>>> - * and drm_syncobj_replace_fence() instead.
>>>>> + * @type: indicate syncobj type
>>>>> */
>>>>> - struct dma_fence __rcu *fence;
>>>>> + enum drm_syncobj_type type;
>>>>> /**
>>>>> - * @cb_list: List of callbacks to call when the &fence gets
>>>>> replaced.
>>>>> + * @syncobj_timeline: timeline
>>>>> */
>>>>> - struct list_head cb_list;
>>>>> + struct drm_syncobj_timeline syncobj_timeline;
>>>>> +
>>>>> /**
>>>>> - * @lock: Protects &cb_list and write-locks &fence.
>>>>> + * @lock: Protects timeline list and write-locks &fence.
>>>>> */
>>>>> spinlock_t lock;
>>>>> /**
>>>>> @@ -62,25 +78,6 @@ struct drm_syncobj {
>>>>> struct file *file;
>>>>> };
>>>>> -typedef void (*drm_syncobj_func_t)(struct drm_syncobj *syncobj,
>>>>> - struct drm_syncobj_cb *cb);
>>>>> -
>>>>> -/**
>>>>> - * struct drm_syncobj_cb - callback for drm_syncobj_add_callback
>>>>> - * @node: used by drm_syncob_add_callback to append this struct to
>>>>> - * &drm_syncobj.cb_list
>>>>> - * @func: drm_syncobj_func_t to call
>>>>> - *
>>>>> - * This struct will be initialized by drm_syncobj_add_callback,
>>>>> additional
>>>>> - * data can be passed along by embedding drm_syncobj_cb in
>>>>> another struct.
>>>>> - * The callback will get called the next time
>>>>> drm_syncobj_replace_fence is
>>>>> - * called.
>>>>> - */
>>>>> -struct drm_syncobj_cb {
>>>>> - struct list_head node;
>>>>> - drm_syncobj_func_t func;
>>>>> -};
>>>>> -
>>>>> void drm_syncobj_free(struct kref *kref);
>>>>> /**
>>>>> @@ -106,29 +103,6 @@ drm_syncobj_put(struct drm_syncobj *obj)
>>>>> kref_put(&obj->refcount, drm_syncobj_free);
>>>>> }
>>>>> -/**
>>>>> - * drm_syncobj_fence_get - get a reference to a fence in a sync
>>>>> object
>>>>> - * @syncobj: sync object.
>>>>> - *
>>>>> - * This acquires additional reference to &drm_syncobj.fence
>>>>> contained in @obj,
>>>>> - * if not NULL. It is illegal to call this without already
>>>>> holding a reference.
>>>>> - * No locks required.
>>>>> - *
>>>>> - * Returns:
>>>>> - * Either the fence of @obj or NULL if there's none.
>>>>> - */
>>>>> -static inline struct dma_fence *
>>>>> -drm_syncobj_fence_get(struct drm_syncobj *syncobj)
>>>>> -{
>>>>> - struct dma_fence *fence;
>>>>> -
>>>>> - rcu_read_lock();
>>>>> - fence = dma_fence_get_rcu_safe(&syncobj->fence);
>>>>> - rcu_read_unlock();
>>>>> -
>>>>> - return fence;
>>>>> -}
>>>>> -
>>>>> struct drm_syncobj *drm_syncobj_find(struct drm_file *file_private,
>>>>> u32 handle);
>>>>> void drm_syncobj_replace_fence(struct drm_syncobj *syncobj, u64
>>>>> point,
>>>>> @@ -142,5 +116,7 @@ int drm_syncobj_create(struct drm_syncobj
>>>>> **out_syncobj, uint32_t flags,
>>>>> int drm_syncobj_get_handle(struct drm_file *file_private,
>>>>> struct drm_syncobj *syncobj, u32 *handle);
>>>>> int drm_syncobj_get_fd(struct drm_syncobj *syncobj, int *p_fd);
>>>>> +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point,
>>>>> + u64 flags, struct dma_fence **fence);
>>>>> #endif
>>>>> diff --git a/include/uapi/drm/drm.h b/include/uapi/drm/drm.h
>>>>> index 300f336633f2..cebdb2541eb7 100644
>>>>> --- a/include/uapi/drm/drm.h
>>>>> +++ b/include/uapi/drm/drm.h
>>>>> @@ -717,6 +717,7 @@ struct drm_prime_handle {
>>>>> struct drm_syncobj_create {
>>>>> __u32 handle;
>>>>> #define DRM_SYNCOBJ_CREATE_SIGNALED (1 << 0)
>>>>> +#define DRM_SYNCOBJ_CREATE_TYPE_TIMELINE (1 << 1)
>>>>> __u32 flags;
>>>>> };
>>>>
>>>
>>> _______________________________________________
>>> amd-gfx mailing list
>>> amd-gfx at lists.freedesktop.org
>>> https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>>
>> _______________________________________________
>> dri-devel mailing list
>> dri-devel at lists.freedesktop.org
>> https://lists.freedesktop.org/mailman/listinfo/dri-devel
>
More information about the amd-gfx
mailing list