[PATCH 1/6] drm/ttm: change fence_lock to inner lock
Thomas Hellstrom
thellstrom at vmware.com
Wed Nov 28 03:40:50 PST 2012
On 11/28/2012 12:25 PM, Maarten Lankhorst wrote:
> This requires changing the order in ttm_bo_cleanup_refs_or_queue to
> take the reservation first, as there is otherwise no race free way to
> take lru lock before fence_lock.
>
> Signed-off-by: Maarten Lankhorst <maarten.lankhorst at canonical.com>
Technically I think it would be possible to avoid reserving in the case
the bo is still busy, and re-checking
for sync_obj presence once we have reserved (since we release the fence
lock).
But that's an optimization. The important thing at this point is to get
this right.
Reviewed-by: Thomas Hellstrom <thellstrom at vmware.com>
> ---
> drivers/gpu/drm/ttm/ttm_bo.c | 31 +++++++++++--------------------
> drivers/gpu/drm/ttm/ttm_execbuf_util.c | 4 ++--
> 2 files changed, 13 insertions(+), 22 deletions(-)
>
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index 7426fe5..202fc20 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -500,27 +500,17 @@ static void ttm_bo_cleanup_refs_or_queue(struct ttm_buffer_object *bo)
> {
> struct ttm_bo_device *bdev = bo->bdev;
> struct ttm_bo_global *glob = bo->glob;
> - struct ttm_bo_driver *driver;
> + struct ttm_bo_driver *driver = bdev->driver;
> void *sync_obj = NULL;
> int put_count;
> int ret;
>
> + spin_lock(&glob->lru_lock);
> + ret = ttm_bo_reserve_locked(bo, false, true, false, 0);
> +
> spin_lock(&bdev->fence_lock);
> (void) ttm_bo_wait(bo, false, false, true);
> - if (!bo->sync_obj) {
> -
> - spin_lock(&glob->lru_lock);
> -
> - /**
> - * Lock inversion between bo:reserve and bdev::fence_lock here,
> - * but that's OK, since we're only trylocking.
> - */
> -
> - ret = ttm_bo_reserve_locked(bo, false, true, false, 0);
> -
> - if (unlikely(ret == -EBUSY))
> - goto queue;
> -
> + if (!ret && !bo->sync_obj) {
> spin_unlock(&bdev->fence_lock);
> put_count = ttm_bo_del_from_lru(bo);
>
> @@ -530,18 +520,19 @@ static void ttm_bo_cleanup_refs_or_queue(struct ttm_buffer_object *bo)
> ttm_bo_list_ref_sub(bo, put_count, true);
>
> return;
> - } else {
> - spin_lock(&glob->lru_lock);
> }
> -queue:
> - driver = bdev->driver;
> if (bo->sync_obj)
> sync_obj = driver->sync_obj_ref(bo->sync_obj);
> + spin_unlock(&bdev->fence_lock);
> +
> + if (!ret) {
> + atomic_set(&bo->reserved, 0);
> + wake_up_all(&bo->event_queue);
> + }
>
> kref_get(&bo->list_kref);
> list_add_tail(&bo->ddestroy, &bdev->ddestroy);
> spin_unlock(&glob->lru_lock);
> - spin_unlock(&bdev->fence_lock);
>
> if (sync_obj) {
> driver->sync_obj_flush(sync_obj);
> diff --git a/drivers/gpu/drm/ttm/ttm_execbuf_util.c b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> index 1986d00..cd9e452 100644
> --- a/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
> @@ -213,8 +213,8 @@ void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj)
> driver = bdev->driver;
> glob = bo->glob;
>
> - spin_lock(&bdev->fence_lock);
> spin_lock(&glob->lru_lock);
> + spin_lock(&bdev->fence_lock);
>
> list_for_each_entry(entry, list, head) {
> bo = entry->bo;
> @@ -223,8 +223,8 @@ void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj)
> ttm_bo_unreserve_locked(bo);
> entry->reserved = false;
> }
> - spin_unlock(&glob->lru_lock);
> spin_unlock(&bdev->fence_lock);
> + spin_unlock(&glob->lru_lock);
>
> list_for_each_entry(entry, list, head) {
> if (entry->old_sync_obj)
More information about the dri-devel
mailing list