[PATCH 10/13] drm/radeon: return -ENOENT in fence_wait_*
Michel Dänzer
michel at daenzer.net
Fri Apr 20 02:15:26 PDT 2012
On Fre, 2012-04-20 at 10:49 +0200, Christian König wrote:
> On 20.04.2012 09:20, Michel Dänzer wrote:
> > On Fre, 2012-04-20 at 00:39 +0200, Christian König wrote:
> >> Signed-off-by: Christian König<deathsimple at vodafone.de>
> >> ---
> >> drivers/gpu/drm/radeon/radeon_fence.c | 4 ++--
> >> 1 files changed, 2 insertions(+), 2 deletions(-)
> >>
> >> diff --git a/drivers/gpu/drm/radeon/radeon_fence.c b/drivers/gpu/drm/radeon/radeon_fence.c
> >> index 1a9765a..764ab7e 100644
> >> --- a/drivers/gpu/drm/radeon/radeon_fence.c
> >> +++ b/drivers/gpu/drm/radeon/radeon_fence.c
> >> @@ -286,7 +286,7 @@ int radeon_fence_wait_next(struct radeon_device *rdev, int ring)
> >> }
> >> if (list_empty(&rdev->fence_drv[ring].emitted)) {
> >> write_unlock_irqrestore(&rdev->fence_lock, irq_flags);
> >> - return 0;
> >> + return -ENOENT;
> >> }
> >> fence = list_entry(rdev->fence_drv[ring].emitted.next,
> >> struct radeon_fence, list);
> >> @@ -310,7 +310,7 @@ int radeon_fence_wait_last(struct radeon_device *rdev, int ring)
> >> }
> >> if (list_empty(&rdev->fence_drv[ring].emitted)) {
> >> write_unlock_irqrestore(&rdev->fence_lock, irq_flags);
> >> - return 0;
> >> + return -ENOENT;
> >> }
> >> fence = list_entry(rdev->fence_drv[ring].emitted.prev,
> >> struct radeon_fence, list);
> > It seems weird to declare a fence wait as failed when there are no
> > outstanding fences in the first place. If there are callers which
> > require outstanding fences, they should probably handle that themselves.
>
> Why that sounds so weird? Ok, maybe for radeon_fence_wait_last that's
> questionable,
Indeed. It happens not to break radeon_suspend_kms because it doesn't
check the return value, but otherwise it would fail spuriously.
> but for radeon_fence_wait_next it's quite clear to me that
> we should signal the caller that there is no fence to wait for.
>
> The problem I wanted to fix with that is the usage of
> radeon_fence_wait_next in radeon_ring_alloc (for example):
> > int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring
> > *ring, unsigned ndw)
> > {
> > int r;
> >
> > /* Align requested size with padding so unlock_commit can
> > * pad safely */
> > ndw = (ndw + ring->align_mask) & ~ring->align_mask;
> > while (ndw > (ring->ring_free_dw - 1)) {
> > radeon_ring_free_size(rdev, ring);
> > if (ndw < ring->ring_free_dw) {
> > break;
> > }
> > r = radeon_fence_wait_next(rdev,
> > radeon_ring_index(rdev, ring));
> > if (r)
> > return r;
> > }
> > ring->count_dw = ndw;
> > ring->wptr_old = ring->wptr;
> > return 0;
> > }
> If the ring is full, but actually has no more fences in it (which in my
> case was caused by my stupidity and actually shouldn't happen otherwise)
> this loop will just busy wait with a critical mutex locked for something
> that never happens.
My suggestion was to explicitly check for that in radeon_ring_alloc. But
I guess right now it doesn't really matter, as it's the only caller. :)
--
Earthling Michel Dänzer | http://www.amd.com
Libre software enthusiast | Debian, X and DRI developer
More information about the dri-devel
mailing list