Include request for reset-rework branch.

Jerome Glisse j.glisse at
Mon Apr 30 09:26:10 PDT 2012

On Mon, Apr 30, 2012 at 11:37 AM, Christian König
<deathsimple at> wrote:
> On 30.04.2012 17:12, Jerome Glisse wrote:
>> On Mon, Apr 30, 2012 at 11:12 AM, Jerome Glisse<j.glisse at>
>>  wrote:
>>> On Mon, Apr 30, 2012 at 10:50 AM, Christian König
>>> <deathsimple at>  wrote:
>>>> Hi Dave,
>>>> if nobody has a last moment concern please include the following patches
>>>> in drm-next.
>>>> Except for some minor fixes they have already been on the list for quite
>>>> some time,
>>>> but I intentional left out the debugfs related patches cause we haven't
>>>> finished the
>>>> discussion about them yet.
>>>> If you prefer to merge them directly, I also made them available as
>>>> reset-rework
>>>> branch here: git://
>>>> Cheers,
>>>> Christian.
>>> I am not completely ok, i am against patch 5. I need sometime to review
>>> it.
>>> Cheers,
>>> Jerome
>> Sorr mean patch 7
> I just started to wonder :) what's wrong with patch 7?
> Please keep in mind that implementing proper locking in the lower level
> objects allows us to remove quite some locking in the upper layers.
> By the way, do you mind when I dig into the whole locking stuff of the
> kernel driver a bit more? There seems to be allot of possibilities to
> cleanup and simplify the overall driver.
> Cheers,
> Christian.

Well when it comes to lock, i have some kind of an idea i wanted to
work for a while. GPU is transaction based in a sense, we feed rings
and it works on them, 90% of the work is cs ioctl so we should be able
to have one and only one lock (ignoring the whole modesetting path
here for which i believe one lock too is enough). Things like PM would
need to take all lock but hey i believe that's a good thing as my
understanding is PM we do right now need most of the GPU idle.

So here is what we have
ih spinlock (can be ignored ie left as is)
irq spinlock (can be ignored ie left as is)
blit mutex
pm mutex
cs mutex
dc_hw_i2c mutex
vram mutex
ib mutex
rings[] lock

So the real issue is ttm calling back into the driver. So the idea i
had is to have a work thread that is the only one allowed to mess with
the GPU. The work thread would use some locking in which it has
preference over the writer (writer being cs ioctl or ttm call back or
anything else that needs to schedule GPU work). This would require
only two lock. I am actually thinking of having 2 list one where
writer add there "transaction" and one where the worker empty the
transaction, something like:

list_add_tail(trans, trans_temp_list)

list_splice_tail(trans_list, trans_temp_list)
// schedule ib
// process fence & semaphore

So there would be one transaction lock and one lock for the
transaction memory allocation (ib, semaphore and alike) The workeur
would also be responsible for GPU reset, there wouldn't be any ring
lock as i believe one worker is enough for all rings.

For transaction the only issue really is ttm, cs is easy it's an
ib+fence+semaphore, ttm can be more complex, it can be bo move, bind,
unbind, ... Anyway it's doable and it's design i had in mind for a


More information about the dri-devel mailing list