[PATCH] drm/ttm/pool: Revert to clear-on-alloc to honor TTM_TT_FLAG_ZERO_ALLOC

Christian König christian.koenig at amd.com
Mon Jun 24 08:41:31 UTC 2024


Am 21.06.24 um 17:43 schrieb Nirmoy Das:
> Hi Christian,
>
> On 6/21/2024 4:54 PM, Christian König wrote:
>> Am 20.06.24 um 18:01 schrieb Nirmoy Das:
>>> Currently ttm pool is not honoring TTM_TT_FLAG_ZERO_ALLOC flag and
>>> clearing pages on free. It does help with allocation latency but 
>>> clearing
>>> happens even if drm driver doesn't passes the flag. If clear on free
>>> is needed then a new flag can be added for that purpose.
>>
>> Mhm, thinking more about it that will most likely get push back from 
>> others as well.
>
>
> Agreed, it is diverting a lot from a known behavior.
>
>>
>> How about the attached patch? We just skip clearing pages when the 
>> driver set the ZERO_ALLOC flag again before freeing them.
>>
>> Maybe rename the flag or add a new one for that, but in general that 
>> looks like the option with the least impact.
>
>
> I would prefer a few flag (TTM_TT_FLAG_CLEARED_ALLOC ?) which driver 
> can set before freeing. I can resend the patch if you are
>
> fine with it.


Yeah, please go ahead with that. But probably better name it 
TTM_TT_FLAG_CLEARED_ON_FREE.

Thanks,
Christian.

>
>
> Regards,
>
> Nirmoy
>
>
>>
>> Regards,
>> Christian.
>>
>>>
>>> Cc: Christian Koenig <christian.koenig at amd.com>
>>> Cc: "Thomas Hellström" <thomas.hellstrom at linux.intel.com>
>>> Cc: Matthew Auld <matthew.auld at intel.com>
>>> Signed-off-by: Nirmoy Das <nirmoy.das at intel.com>
>>> ---
>>>   drivers/gpu/drm/ttm/ttm_pool.c | 31 +++++++++++++++++--------------
>>>   1 file changed, 17 insertions(+), 14 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/ttm/ttm_pool.c 
>>> b/drivers/gpu/drm/ttm/ttm_pool.c
>>> index 6e1fd6985ffc..cbbd722185ee 100644
>>> --- a/drivers/gpu/drm/ttm/ttm_pool.c
>>> +++ b/drivers/gpu/drm/ttm/ttm_pool.c
>>> @@ -224,15 +224,6 @@ static void ttm_pool_unmap(struct ttm_pool 
>>> *pool, dma_addr_t dma_addr,
>>>   /* Give pages into a specific pool_type */
>>>   static void ttm_pool_type_give(struct ttm_pool_type *pt, struct 
>>> page *p)
>>>   {
>>> -    unsigned int i, num_pages = 1 << pt->order;
>>> -
>>> -    for (i = 0; i < num_pages; ++i) {
>>> -        if (PageHighMem(p))
>>> -            clear_highpage(p + i);
>>> -        else
>>> -            clear_page(page_address(p + i));
>>> -    }
>>> -
>>>       spin_lock(&pt->lock);
>>>       list_add(&p->lru, &pt->pages);
>>>       spin_unlock(&pt->lock);
>>> @@ -240,15 +231,26 @@ static void ttm_pool_type_give(struct 
>>> ttm_pool_type *pt, struct page *p)
>>>   }
>>>     /* Take pages from a specific pool_type, return NULL when 
>>> nothing available */
>>> -static struct page *ttm_pool_type_take(struct ttm_pool_type *pt)
>>> +static struct page *ttm_pool_type_take(struct ttm_pool_type *pt, 
>>> bool clear)
>>>   {
>>>       struct page *p;
>>>         spin_lock(&pt->lock);
>>>       p = list_first_entry_or_null(&pt->pages, typeof(*p), lru);
>>>       if (p) {
>>> +        unsigned int i, num_pages = 1 << pt->order;
>>> +
>>>           atomic_long_sub(1 << pt->order, &allocated_pages);
>>>           list_del(&p->lru);
>>> +        if (clear) {
>>> +            for (i = 0; i < num_pages; ++i) {
>>> +                if (PageHighMem(p))
>>> +                    clear_highpage(p + i);
>>> +                else
>>> +                    clear_page(page_address(p + i));
>>> +            }
>>> +        }
>>> +
>>>       }
>>>       spin_unlock(&pt->lock);
>>>   @@ -279,7 +281,7 @@ static void ttm_pool_type_fini(struct 
>>> ttm_pool_type *pt)
>>>       list_del(&pt->shrinker_list);
>>>       spin_unlock(&shrinker_lock);
>>>   -    while ((p = ttm_pool_type_take(pt)))
>>> +    while ((p = ttm_pool_type_take(pt, false)))
>>>           ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
>>>   }
>>>   @@ -330,7 +332,7 @@ static unsigned int ttm_pool_shrink(void)
>>>       list_move_tail(&pt->shrinker_list, &shrinker_list);
>>>       spin_unlock(&shrinker_lock);
>>>   -    p = ttm_pool_type_take(pt);
>>> +    p = ttm_pool_type_take(pt, false);
>>>       if (p) {
>>>           ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
>>>           num_pages = 1 << pt->order;
>>> @@ -457,10 +459,11 @@ int ttm_pool_alloc(struct ttm_pool *pool, 
>>> struct ttm_tt *tt,
>>>            num_pages;
>>>            order = min_t(unsigned int, order, __fls(num_pages))) {
>>>           struct ttm_pool_type *pt;
>>> +        bool clear = tt->page_flags & TTM_TT_FLAG_ZERO_ALLOC;
>>>             page_caching = tt->caching;
>>>           pt = ttm_pool_select_type(pool, tt->caching, order);
>>> -        p = pt ? ttm_pool_type_take(pt) : NULL;
>>> +        p = pt ? ttm_pool_type_take(pt, clear) : NULL;
>>>           if (p) {
>>>               r = ttm_pool_apply_caching(caching, pages,
>>>                              tt->caching);
>>> @@ -480,7 +483,7 @@ int ttm_pool_alloc(struct ttm_pool *pool, struct 
>>> ttm_tt *tt,
>>>                   if (num_pages < (1 << order))
>>>                       break;
>>>   -                p = ttm_pool_type_take(pt);
>>> +                p = ttm_pool_type_take(pt, clear);
>>>               } while (p);
>>>           }



More information about the dri-devel mailing list