[PATCH 2/2] drm/ttm: completely rework ttm_bo_delayed_delete
Thomas Hellstrom
thomas at shipmail.org
Wed Dec 13 20:55:28 UTC 2017
Hi, Christian,
While this has probably already been committed, and looks like a nice
cleanup there are two things below I think needs fixing.
On 11/15/2017 01:31 PM, Christian König wrote:
> There is no guarantee that the next entry on the ddelete list stays on
> the list when we drop the locks.
>
> Completely rework this mess by moving processed entries on a temporary
> list.
>
> Signed-off-by: Christian König <christian.koenig at amd.com>
> ---
> drivers/gpu/drm/ttm/ttm_bo.c | 77 ++++++++++++++------------------------------
> 1 file changed, 25 insertions(+), 52 deletions(-)
>
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index 7c1eac4f4b4b..ad0afdd71f21 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -572,71 +572,47 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo,
> * Traverse the delayed list, and call ttm_bo_cleanup_refs on all
> * encountered buffers.
> */
> -
> -static int ttm_bo_delayed_delete(struct ttm_bo_device *bdev, bool remove_all)
> +static bool ttm_bo_delayed_delete(struct ttm_bo_device *bdev, bool remove_all)
> {
> struct ttm_bo_global *glob = bdev->glob;
> - struct ttm_buffer_object *entry = NULL;
> - int ret = 0;
> -
> - spin_lock(&glob->lru_lock);
> - if (list_empty(&bdev->ddestroy))
> - goto out_unlock;
> + struct list_head removed;
> + bool empty;
>
> - entry = list_first_entry(&bdev->ddestroy,
> - struct ttm_buffer_object, ddestroy);
> - kref_get(&entry->list_kref);
> + INIT_LIST_HEAD(&removed);
>
> - for (;;) {
> - struct ttm_buffer_object *nentry = NULL;
> -
> - if (entry->ddestroy.next != &bdev->ddestroy) {
> - nentry = list_first_entry(&entry->ddestroy,
> - struct ttm_buffer_object, ddestroy);
> - kref_get(&nentry->list_kref);
> - }
> -
> - ret = reservation_object_trylock(entry->resv) ? 0 : -EBUSY;
> - if (remove_all && ret) {
> - spin_unlock(&glob->lru_lock);
> - ret = reservation_object_lock(entry->resv, NULL);
> - spin_lock(&glob->lru_lock);
> - }
> + spin_lock(&glob->lru_lock);
> + while (!list_empty(&bdev->ddestroy)) {
> + struct ttm_buffer_object *bo;
>
> - if (!ret)
> - ret = ttm_bo_cleanup_refs(entry, false, !remove_all,
> - true);
> - else
> - spin_unlock(&glob->lru_lock);
> + bo = list_first_entry(&bdev->ddestroy, struct ttm_buffer_object,
> + ddestroy);
> + kref_get(&bo->list_kref);
> + list_move_tail(&bo->ddestroy, &removed);
> + spin_unlock(&glob->lru_lock);
>
> - kref_put(&entry->list_kref, ttm_bo_release_list);
> - entry = nentry;
> + reservation_object_lock(bo->resv, NULL);
Reservation may be a long lived lock, and typically if the object is
reserved here, it's being evicted somewhere and there might be a
substantial stall, which isn't really acceptable in the global
workqueue. Better to move on to the next bo.
This function was really intended to be non-blocking, unless remove_all
== true. I even think it's safe to keep the spinlock held on tryreserve?
>
> - if (ret || !entry)
> - goto out;
> + spin_lock(&glob->lru_lock);
> + ttm_bo_cleanup_refs(bo, false, !remove_all, true);
>
> + kref_put(&bo->list_kref, ttm_bo_release_list);
Calling a release function in atomic context is a bad thing. Nobody
knows what locks needs to be taken in the release function and such code
is prone to lock inversion and sleep-while-atomic bugs. Not long ago
vfree() was even forbidden from atomic context. But here it's easily
avoidable.
/Thomas
More information about the dri-devel
mailing list